[JENKINS] Lucene-Solr-5.x-Linux (64bit/jdk1.9.0-ea-b60) - Build # 12665 - Failure!

2015-05-27 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-Linux/12665/
Java: 64bit/jdk1.9.0-ea-b60 -XX:-UseCompressedOops -XX:+UseSerialGC

1 tests failed.
FAILED:  org.apache.solr.cloud.MultiThreadedOCPTest.test

Error Message:
Captured an uncaught exception in thread: Thread[id=2619, 
name=parallelCoreAdminExecutor-1404-thread-7, state=RUNNABLE, 
group=TGRP-MultiThreadedOCPTest]

Stack Trace:
com.carrotsearch.randomizedtesting.UncaughtExceptionError: Captured an uncaught 
exception in thread: Thread[id=2619, 
name=parallelCoreAdminExecutor-1404-thread-7, state=RUNNABLE, 
group=TGRP-MultiThreadedOCPTest]
at 
__randomizedtesting.SeedInfo.seed([8C5E7FB8C8AB60AD:40A406266570D55]:0)
Caused by: java.lang.AssertionError: Too many closes on SolrCore
at __randomizedtesting.SeedInfo.seed([8C5E7FB8C8AB60AD]:0)
at org.apache.solr.core.SolrCore.close(SolrCore.java:1150)
at org.apache.solr.common.util.IOUtils.closeQuietly(IOUtils.java:31)
at org.apache.solr.core.CoreContainer.create(CoreContainer.java:652)
at org.apache.solr.core.CoreContainer.create(CoreContainer.java:611)
at 
org.apache.solr.handler.admin.CoreAdminHandler.handleCreateAction(CoreAdminHandler.java:628)
at 
org.apache.solr.handler.admin.CoreAdminHandler.handleRequestInternal(CoreAdminHandler.java:213)
at 
org.apache.solr.handler.admin.CoreAdminHandler$ParallelCoreAdminHandlerThread.run(CoreAdminHandler.java:1249)
at 
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor$1.run(ExecutorUtil.java:148)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)




Build Log:
[...truncated 10034 lines...]
   [junit4] Suite: org.apache.solr.cloud.MultiThreadedOCPTest
   [junit4]   2 Creating dataDir: 
/home/jenkins/workspace/Lucene-Solr-5.x-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.MultiThreadedOCPTest
 8C5E7FB8C8AB60AD-001/init-core-data-001
   [junit4]   2 246209 T2279 oas.SolrTestCaseJ4.buildSSLConfig Randomized ssl 
(false) and clientAuth (true)
   [junit4]   2 246209 T2279 oas.BaseDistributedSearchTestCase.initHostContext 
Setting hostContext system property: /gndv/a
   [junit4]   2 246211 T2279 oasc.ZkTestServer.run STARTING ZK TEST SERVER
   [junit4]   2 246211 T2280 oasc.ZkTestServer$2$1.setClientPort client 
port:0.0.0.0/0.0.0.0:0
   [junit4]   2 246211 T2280 oasc.ZkTestServer$ZKServerMain.runFromConfig 
Starting server
   [junit4]   2 246311 T2279 oasc.ZkTestServer.run start zk server on 
port:59109
   [junit4]   2 246312 T2279 
oascc.SolrZkClient.createZkCredentialsToAddAutomatically Using default 
ZkCredentialsProvider
   [junit4]   2 246313 T2279 oascc.ConnectionManager.waitForConnected Waiting 
for client to connect to ZooKeeper
   [junit4]   2 246316 T2287 oascc.ConnectionManager.process Watcher 
org.apache.solr.common.cloud.ConnectionManager@293a7a58 
name:ZooKeeperConnection Watcher:127.0.0.1:59109 got event WatchedEvent 
state:SyncConnected type:None path:null path:null type:None
   [junit4]   2 246316 T2279 oascc.ConnectionManager.waitForConnected Client 
is connected to ZooKeeper
   [junit4]   2 246316 T2279 oascc.SolrZkClient.createZkACLProvider Using 
default ZkACLProvider
   [junit4]   2 246317 T2279 oascc.SolrZkClient.makePath makePath: /solr
   [junit4]   2 246318 T2279 
oascc.SolrZkClient.createZkCredentialsToAddAutomatically Using default 
ZkCredentialsProvider
   [junit4]   2 246319 T2279 oascc.ConnectionManager.waitForConnected Waiting 
for client to connect to ZooKeeper
   [junit4]   2 246335 T2290 oascc.ConnectionManager.process Watcher 
org.apache.solr.common.cloud.ConnectionManager@5a976749 
name:ZooKeeperConnection Watcher:127.0.0.1:59109/solr got event WatchedEvent 
state:SyncConnected type:None path:null path:null type:None
   [junit4]   2 246335 T2279 oascc.ConnectionManager.waitForConnected Client 
is connected to ZooKeeper
   [junit4]   2 246335 T2279 oascc.SolrZkClient.createZkACLProvider Using 
default ZkACLProvider
   [junit4]   2 246336 T2279 oascc.SolrZkClient.makePath makePath: 
/collections/collection1
   [junit4]   2 246338 T2279 oascc.SolrZkClient.makePath makePath: 
/collections/collection1/shards
   [junit4]   2 246339 T2279 oascc.SolrZkClient.makePath makePath: 
/collections/control_collection
   [junit4]   2 246340 T2279 oascc.SolrZkClient.makePath makePath: 
/collections/control_collection/shards
   [junit4]   2 246341 T2279 oasc.AbstractZkTestCase.putConfig put 
/home/jenkins/workspace/Lucene-Solr-5.x-Linux/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml
 to /configs/conf1/solrconfig.xml
   [junit4]   2 246342 T2279 oascc.SolrZkClient.makePath makePath: 
/configs/conf1/solrconfig.xml
   [junit4]   2 246343 T2279 oasc.AbstractZkTestCase.putConfig put 

[jira] [Commented] (LUCENE-6481) Improve GeoPointField type to only visit high precision boundary terms

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6481?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560693#comment-14560693
 ] 

ASF subversion and git services commented on LUCENE-6481:
-

Commit 1681942 from [~mikemccand] in branch 'dev/branches/LUCENE-6481'
[ https://svn.apache.org/r1681942 ]

LUCENE-6481: add nocommit

 Improve GeoPointField type to only visit high precision boundary terms 
 ---

 Key: LUCENE-6481
 URL: https://issues.apache.org/jira/browse/LUCENE-6481
 Project: Lucene - Core
  Issue Type: Improvement
  Components: core/index
Reporter: Nicholas Knize
 Attachments: LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481.patch, 
 LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481_WIP.patch


 Current GeoPointField [LUCENE-6450 | 
 https://issues.apache.org/jira/browse/LUCENE-6450] computes a set of ranges 
 along the space-filling curve that represent a provided bounding box.  This 
 determines which terms to visit in the terms dictionary and which to skip. 
 This is suboptimal for large bounding boxes as we may end up visiting all 
 terms (which could be quite large). 
 This incremental improvement is to improve GeoPointField to only visit high 
 precision terms in boundary ranges and use the postings list for ranges that 
 are completely within the target bounding box.
 A separate improvement is to switch over to auto-prefix and build an 
 Automaton representing the bounding box.  That can be tracked in a separate 
 issue.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6371) Improve Spans payload collection

2015-05-27 Thread Alan Woodward (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6371?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560771#comment-14560771
 ] 

Alan Woodward commented on LUCENE-6371:
---

I'm not sure I'm following what you mean by having two APIs.  
SpanQuery.createWeight() has exactly the same signature as Query.createWeight() 
- replacing needsScores with the termcontexts map in the SpanWeight constructor 
just allows us to pass in the needed information to build the Similarity at the 
same time as indicating whether or not it should be built at all.

 Improve Spans payload collection
 

 Key: LUCENE-6371
 URL: https://issues.apache.org/jira/browse/LUCENE-6371
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Paul Elschot
Assignee: Alan Woodward
Priority: Minor
 Fix For: Trunk, 5.3

 Attachments: LUCENE-6371.patch, LUCENE-6371.patch, LUCENE-6371.patch, 
 LUCENE-6371.patch


 Spin off from LUCENE-6308, see the comments there from around 23 March 2015.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6371) Improve Spans payload collection

2015-05-27 Thread Alan Woodward (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6371?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560783#comment-14560783
 ] 

Alan Woodward commented on LUCENE-6371:
---

Ah, ok.  This is more to do with LUCENE-6466.  Moving .getSpans() and 
.extractTerms() to SpanWeight means that we can't collect the termcontexts in 
the constructor any more, because we can't call them on a partially constructed 
object.  And the TermContexts map was already in the API, but it was in 
getSpans() rather than in the SpanWeight constructor.  You're right that it's 
messy though.  I'm just not sure there's a cleaner way of doing it that doesn't 
involve completely changing how the SpanScorer works.

 Improve Spans payload collection
 

 Key: LUCENE-6371
 URL: https://issues.apache.org/jira/browse/LUCENE-6371
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Paul Elschot
Assignee: Alan Woodward
Priority: Minor
 Fix For: Trunk, 5.3

 Attachments: LUCENE-6371.patch, LUCENE-6371.patch, LUCENE-6371.patch, 
 LUCENE-6371.patch


 Spin off from LUCENE-6308, see the comments there from around 23 March 2015.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6481) Improve GeoPointField type to only visit high precision boundary terms

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6481?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560684#comment-14560684
 ] 

ASF subversion and git services commented on LUCENE-6481:
-

Commit 1681941 from [~mikemccand] in branch 'dev/branches/LUCENE-6481'
[ https://svn.apache.org/r1681941 ]

LUCENE-6481: sometimes use smallish bbox for random test

 Improve GeoPointField type to only visit high precision boundary terms 
 ---

 Key: LUCENE-6481
 URL: https://issues.apache.org/jira/browse/LUCENE-6481
 Project: Lucene - Core
  Issue Type: Improvement
  Components: core/index
Reporter: Nicholas Knize
 Attachments: LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481.patch, 
 LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481_WIP.patch


 Current GeoPointField [LUCENE-6450 | 
 https://issues.apache.org/jira/browse/LUCENE-6450] computes a set of ranges 
 along the space-filling curve that represent a provided bounding box.  This 
 determines which terms to visit in the terms dictionary and which to skip. 
 This is suboptimal for large bounding boxes as we may end up visiting all 
 terms (which could be quite large). 
 This incremental improvement is to improve GeoPointField to only visit high 
 precision terms in boundary ranges and use the postings list for ranges that 
 are completely within the target bounding box.
 A separate improvement is to switch over to auto-prefix and build an 
 Automaton representing the bounding box.  That can be tracked in a separate 
 issue.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6481) Improve GeoPointField type to only visit high precision boundary terms

2015-05-27 Thread Michael McCandless (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6481?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560690#comment-14560690
 ] 

Michael McCandless commented on LUCENE-6481:


I committed a small improvement to the test, to sometimes test on a smallish 
random sized bbox, and it hits this new failure:

{noformat}
   [junit4] Started J0 PID(15031@localhost).
   [junit4] Suite: org.apache.lucene.search.TestGeoPointQuery
   [junit4]   1 smallBBox=true
   [junit4]   1 T0: iter=1 id=3 docID=3 lat=-90.2957562284163 
lon=-179.83102465432302 (bbox: lat=-90.84003988710097 TO -89.86198542441215 
lon=-180.5339936365967 TO -179.31722107698667) expected true but got: false 
deleted?=false query=GeoPointInBBoxQuery: field=geoField: Lower Left: 
[-180.5339936365967,-90.84003988710097] Upper Right: 
[-179.31722107698667,-89.86198542441215]
   [junit4]   2 ??? 27, 2015 12:37:04 PM 
com.carrotsearch.randomizedtesting.RandomizedRunner$QueueUncaughtExceptionsHandler
 uncaughtException
   [junit4]   2 WARNING: Uncaught exception in thread: 
Thread[T0,5,TGRP-TestGeoPointQuery]
   [junit4]   2 java.lang.AssertionError: wrong result
   [junit4]   2at __randomizedtesting.SeedInfo.seed([5]:0)
   [junit4]   2at org.junit.Assert.fail(Assert.java:93)
   [junit4]   2at 
org.apache.lucene.search.TestGeoPointQuery$1._run(TestGeoPointQuery.java:372)
   [junit4]   2at 
org.apache.lucene.search.TestGeoPointQuery$1.run(TestGeoPointQuery.java:271)
   [junit4]   2 
   [junit4]   2 NOTE: reproduce with: ant test  -Dtestcase=TestGeoPointQuery 
-Dtests.method=testRandomTiny -Dtests.seed=5 -Dtests.locale=sr_ME 
-Dtests.timezone=Europe/Vilnius -Dtests.asserts=true 
-Dtests.file.encoding=US-ASCII
   [junit4] ERROR   0.09s | TestGeoPointQuery.testRandomTiny 
{noformat}

I'm not sure why testRandomTiny is hitting all these problems!

 Improve GeoPointField type to only visit high precision boundary terms 
 ---

 Key: LUCENE-6481
 URL: https://issues.apache.org/jira/browse/LUCENE-6481
 Project: Lucene - Core
  Issue Type: Improvement
  Components: core/index
Reporter: Nicholas Knize
 Attachments: LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481.patch, 
 LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481_WIP.patch


 Current GeoPointField [LUCENE-6450 | 
 https://issues.apache.org/jira/browse/LUCENE-6450] computes a set of ranges 
 along the space-filling curve that represent a provided bounding box.  This 
 determines which terms to visit in the terms dictionary and which to skip. 
 This is suboptimal for large bounding boxes as we may end up visiting all 
 terms (which could be quite large). 
 This incremental improvement is to improve GeoPointField to only visit high 
 precision terms in boundary ranges and use the postings list for ranges that 
 are completely within the target bounding box.
 A separate improvement is to switch over to auto-prefix and build an 
 Automaton representing the bounding box.  That can be tracked in a separate 
 issue.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6481) Improve GeoPointField type to only visit high precision boundary terms

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6481?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560702#comment-14560702
 ] 

ASF subversion and git services commented on LUCENE-6481:
-

Commit 1681943 from [~mikemccand] in branch 'dev/branches/LUCENE-6481'
[ https://svn.apache.org/r1681943 ]

LUCENE-6481: woops

 Improve GeoPointField type to only visit high precision boundary terms 
 ---

 Key: LUCENE-6481
 URL: https://issues.apache.org/jira/browse/LUCENE-6481
 Project: Lucene - Core
  Issue Type: Improvement
  Components: core/index
Reporter: Nicholas Knize
 Attachments: LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481.patch, 
 LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481_WIP.patch


 Current GeoPointField [LUCENE-6450 | 
 https://issues.apache.org/jira/browse/LUCENE-6450] computes a set of ranges 
 along the space-filling curve that represent a provided bounding box.  This 
 determines which terms to visit in the terms dictionary and which to skip. 
 This is suboptimal for large bounding boxes as we may end up visiting all 
 terms (which could be quite large). 
 This incremental improvement is to improve GeoPointField to only visit high 
 precision terms in boundary ranges and use the postings list for ranges that 
 are completely within the target bounding box.
 A separate improvement is to switch over to auto-prefix and build an 
 Automaton representing the bounding box.  That can be tracked in a separate 
 issue.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-trunk-Windows (32bit/jdk1.8.0_45) - Build # 4859 - Failure!

2015-05-27 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-Windows/4859/
Java: 32bit/jdk1.8.0_45 -server -XX:+UseG1GC

1 tests failed.
FAILED:  org.apache.solr.cloud.MultiThreadedOCPTest.test

Error Message:
Captured an uncaught exception in thread: Thread[id=9599, 
name=parallelCoreAdminExecutor-4953-thread-15, state=RUNNABLE, 
group=TGRP-MultiThreadedOCPTest]

Stack Trace:
com.carrotsearch.randomizedtesting.UncaughtExceptionError: Captured an uncaught 
exception in thread: Thread[id=9599, 
name=parallelCoreAdminExecutor-4953-thread-15, state=RUNNABLE, 
group=TGRP-MultiThreadedOCPTest]
at 
__randomizedtesting.SeedInfo.seed([FF25C44C4CBB8E2A:7771FB96E247E3D2]:0)
Caused by: java.lang.AssertionError: Too many closes on SolrCore
at __randomizedtesting.SeedInfo.seed([FF25C44C4CBB8E2A]:0)
at org.apache.solr.core.SolrCore.close(SolrCore.java:1150)
at org.apache.solr.common.util.IOUtils.closeQuietly(IOUtils.java:31)
at org.apache.solr.core.CoreContainer.create(CoreContainer.java:652)
at org.apache.solr.core.CoreContainer.create(CoreContainer.java:611)
at 
org.apache.solr.handler.admin.CoreAdminHandler.handleCreateAction(CoreAdminHandler.java:628)
at 
org.apache.solr.handler.admin.CoreAdminHandler.handleRequestInternal(CoreAdminHandler.java:213)
at 
org.apache.solr.handler.admin.CoreAdminHandler$ParallelCoreAdminHandlerThread.run(CoreAdminHandler.java:1249)
at 
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor$1.run(ExecutorUtil.java:148)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)




Build Log:
[...truncated 10605 lines...]
   [junit4] Suite: org.apache.solr.cloud.MultiThreadedOCPTest
   [junit4]   2 Creating dataDir: 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build\solr-core\test\J1\temp\solr.cloud.MultiThreadedOCPTest
 FF25C44C4CBB8E2A-001\init-core-data-001
   [junit4]   2 1768625 T9263 oas.SolrTestCaseJ4.buildSSLConfig Randomized ssl 
(false) and clientAuth (true)
   [junit4]   2 1768626 T9263 
oas.BaseDistributedSearchTestCase.initHostContext Setting hostContext system 
property: /
   [junit4]   2 1768627 T9263 oasc.ZkTestServer.run STARTING ZK TEST SERVER
   [junit4]   2 1768628 T9264 oasc.ZkTestServer$2$1.setClientPort client 
port:0.0.0.0/0.0.0.0:0
   [junit4]   2 1768629 T9264 oasc.ZkTestServer$ZKServerMain.runFromConfig 
Starting server
   [junit4]   2 1768701 T9263 oasc.ZkTestServer.run start zk server on 
port:57319
   [junit4]   2 1768729 T9263 oasc.AbstractZkTestCase.putConfig put 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\core\src\test-files\solr\collection1\conf\solrconfig-tlog.xml
 to /configs/conf1/solrconfig.xml
   [junit4]   2 1768737 T9263 oasc.AbstractZkTestCase.putConfig put 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\core\src\test-files\solr\collection1\conf\schema.xml
 to /configs/conf1/schema.xml
   [junit4]   2 1768742 T9263 oasc.AbstractZkTestCase.putConfig put 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\core\src\test-files\solr\collection1\conf\solrconfig.snippet.randomindexconfig.xml
 to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
   [junit4]   2 1768746 T9263 oasc.AbstractZkTestCase.putConfig put 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\core\src\test-files\solr\collection1\conf\stopwords.txt
 to /configs/conf1/stopwords.txt
   [junit4]   2 1768749 T9263 oasc.AbstractZkTestCase.putConfig put 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\core\src\test-files\solr\collection1\conf\protwords.txt
 to /configs/conf1/protwords.txt
   [junit4]   2 1768753 T9263 oasc.AbstractZkTestCase.putConfig put 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\core\src\test-files\solr\collection1\conf\currency.xml
 to /configs/conf1/currency.xml
   [junit4]   2 1768759 T9263 oasc.AbstractZkTestCase.putConfig put 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\core\src\test-files\solr\collection1\conf\enumsConfig.xml
 to /configs/conf1/enumsConfig.xml
   [junit4]   2 1768764 T9263 oasc.AbstractZkTestCase.putConfig put 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\core\src\test-files\solr\collection1\conf\open-exchange-rates.json
 to /configs/conf1/open-exchange-rates.json
   [junit4]   2 1768770 T9263 oasc.AbstractZkTestCase.putConfig put 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\core\src\test-files\solr\collection1\conf\mapping-ISOLatin1Accent.txt
 to /configs/conf1/mapping-ISOLatin1Accent.txt
   [junit4]   2 1768775 T9263 oasc.AbstractZkTestCase.putConfig put 
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\core\src\test-files\solr\collection1\conf\old_synonyms.txt
 to /configs/conf1/old_synonyms.txt
   

[jira] [Commented] (LUCENE-6371) Improve Spans payload collection

2015-05-27 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6371?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560775#comment-14560775
 ] 

Robert Muir commented on LUCENE-6371:
-

Sorry, I am not very clear. I don't understand why SpanWeight needs to take 
nulls, or TermContext stuff at all.

Why can't it work like today? If someone Span* impl needs to know when scoring 
is not needed, add {{boolean needsScores}} to SpanWeight's ctor, rather than 
some magic null value. TermContexts map is an impl detail, not an API. Lets not 
let it become one, its too messy for that.

 Improve Spans payload collection
 

 Key: LUCENE-6371
 URL: https://issues.apache.org/jira/browse/LUCENE-6371
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Paul Elschot
Assignee: Alan Woodward
Priority: Minor
 Fix For: Trunk, 5.3

 Attachments: LUCENE-6371.patch, LUCENE-6371.patch, LUCENE-6371.patch, 
 LUCENE-6371.patch


 Spin off from LUCENE-6308, see the comments there from around 23 March 2015.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-7183) SaslZkACLProviderTest reproducible failures due to poor locale blacklisting

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7183?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560610#comment-14560610
 ] 

ASF subversion and git services commented on SOLR-7183:
---

Commit 1681928 from [~anshumg] in branch 'dev/branches/lucene_solr_5_2'
[ https://svn.apache.org/r1681928 ]

SOLR-7183: Fix Locale blacklisting for Minikdc based tests.(merge from 
branch_5x)

 SaslZkACLProviderTest reproducible failures due to poor locale blacklisting
 ---

 Key: SOLR-7183
 URL: https://issues.apache.org/jira/browse/SOLR-7183
 Project: Solr
  Issue Type: Bug
Reporter: Hoss Man
Assignee: Gregory Chanan
 Fix For: 5.2

 Attachments: SOLR-7183.patch


 SaslZkACLProviderTest has this blacklist of locales...
 {code}
   // These Locales don't generate dates that are compatibile with Hadoop 
 MiniKdc.
   protected final static ListString brokenLocales =
 Arrays.asList(
   th_TH_TH_#u-nu-thai,
   ja_JP_JP_#u-ca-japanese,
   hi_IN);
 {code}
 ..but this list is incomplete -- notably because it only focuses on one 
 specific Thai variant, and then does a string Locale.toString() comparison.  
 so at a minimum {{-Dtests.locale=th_TH}} also fails - i suspect there are 
 other variants that will fail as well
 * if there is a bug in Hadoop MiniKdc then that bug should be filed in 
 jira, and there should be Solr jira that refers to it -- the Solr jira URL 
 needs to be included her in the test case so developers in the future can 
 understand the context and have some idea of if/when the third-party lib bug 
 is fixed
 * if we need to work around some Locales because of this bug, then Locale 
 comparisons need be based on whatever aspects of the Locale are actually 
 problematic
 see for example SOLR-6387  this commit: 
 https://svn.apache.org/viewvc/lucene/dev/branches/branch_4x/solr/contrib/morphlines-core/src/test/org/apache/solr/morphlines/solr/AbstractSolrMorphlineZkTestBase.java?r1=1618676r2=1618675pathrev=1618676
 Or SOLR-6991 + TIKA-1526  this commit: 
 https://svn.apache.org/viewvc/lucene/dev/branches/lucene_solr_5_0/solr/contrib/extraction/src/test/org/apache/solr/handler/extraction/ExtractingRequestHandlerTest.java?r1=1653708r2=1653707pathrev=1653708



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-5.x-Linux (64bit/jdk1.8.0_45) - Build # 12644 - Failure!

2015-05-27 Thread Anshum Gupta
8-real-core Haswell-E with 64G DDR4 memory and a NVMe 750-series SSD.

Can run *all* of the Lucene and Solr tests in 10 minutes by running
multiple ant jobs in parallel!

On Wed, May 27, 2015 at 1:17 AM, Ramkumar R. Aiyengar 
andyetitmo...@gmail.com wrote:

 Curious.. sarowe, what's the spec?
 On 26 May 2015 20:41, Anshum Gupta ans...@anshumgupta.net wrote:

 The last buch of fixes seems to have fixed this. The tests passed on a
 Jenkins that had failing tests earlier.
 Thanks Steve Rowe for lending the super-powerful machine that runs the
 entire suite in 8 min!

 I've seen about 20 runs on that box and also runs on Policeman Jenkins
 with no issues related to this test since the last commit so I've also
 back-ported this to 5x as well.

 On Tue, May 26, 2015 at 9:19 AM, Chris Hostetter 
 hossman_luc...@fucit.org wrote:


 : Right, as I said, we weren't hitting this issue due to our Kerberos
 conf.
 : file. i.e. the only thing that was different on our machines as
 compared to
 : everyone else and moving that conf file got the tests to fail for me.
 It
 : now fails fairly consistently without the patch (from SOLR-7468) and
 has
 : been looking good with the patch.

 that smells like the kind of thing that sould have some assume sanity
 checks built into it.

 Given:
 * the test setups a special env before running the test
 * the test assumes the specific env will exist
 * the user's machine may already have env properties set before running
 ant that affect the expected special env

 therefore: before the test does the setup of the special env, it should
 sanity check that the users basic env doesn't have any properties that
 violate the basic situation.

 so, hypothetical example based on what little i understand the
 authentication stuff: if the purpose of a test is to prove that some
 requests work with (or w/o) kerberos authentication, then before doing
 any
 setup of a mock kerberos env (or before setting up a mock situation
 where no authentication is required), the test should verify that there
 isn't already an existing kerberos env. (or if possible: unset whatever
 env/properties define that env)


 trivial example of a similar situation is the script engine tests --
 TestBadConfigs.testBogusScriptEngine:  the purpose of the test is to
 ensure that a solrconfig.xml file that refers to a script engine (by
 name) which is not installed on the machine will produce an expeted error
 at Solr init.  before doing the Solr init, we have an whitebox assume
 that
 asks the JVM directly if a script engine with that name already exists)



 -Hoss
 http://www.lucidworks.com/

 -
 To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
 For additional commands, e-mail: dev-h...@lucene.apache.org




 --
 Anshum Gupta




-- 
Anshum Gupta


[jira] [Commented] (LUCENE-6481) Improve GeoPointField type to only visit high precision boundary terms

2015-05-27 Thread Michael McCandless (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6481?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560704#comment-14560704
 ] 

Michael McCandless commented on LUCENE-6481:


bq. I committed a small improvement to the test, to sometimes test on a 
smallish random sized bbox, and it hits this new failure:

Woops, my bad: test bug.  I committed a fix...

 Improve GeoPointField type to only visit high precision boundary terms 
 ---

 Key: LUCENE-6481
 URL: https://issues.apache.org/jira/browse/LUCENE-6481
 Project: Lucene - Core
  Issue Type: Improvement
  Components: core/index
Reporter: Nicholas Knize
 Attachments: LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481.patch, 
 LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481_WIP.patch


 Current GeoPointField [LUCENE-6450 | 
 https://issues.apache.org/jira/browse/LUCENE-6450] computes a set of ranges 
 along the space-filling curve that represent a provided bounding box.  This 
 determines which terms to visit in the terms dictionary and which to skip. 
 This is suboptimal for large bounding boxes as we may end up visiting all 
 terms (which could be quite large). 
 This incremental improvement is to improve GeoPointField to only visit high 
 precision terms in boundary ranges and use the postings list for ranges that 
 are completely within the target bounding box.
 A separate improvement is to switch over to auto-prefix and build an 
 Automaton representing the bounding box.  That can be tracked in a separate 
 issue.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-5.x-Linux (64bit/jdk1.8.0_45) - Build # 12644 - Failure!

2015-05-27 Thread Ramkumar R. Aiyengar
Curious.. sarowe, what's the spec?
On 26 May 2015 20:41, Anshum Gupta ans...@anshumgupta.net wrote:

 The last buch of fixes seems to have fixed this. The tests passed on a
 Jenkins that had failing tests earlier.
 Thanks Steve Rowe for lending the super-powerful machine that runs the
 entire suite in 8 min!

 I've seen about 20 runs on that box and also runs on Policeman Jenkins
 with no issues related to this test since the last commit so I've also
 back-ported this to 5x as well.

 On Tue, May 26, 2015 at 9:19 AM, Chris Hostetter hossman_luc...@fucit.org
  wrote:


 : Right, as I said, we weren't hitting this issue due to our Kerberos
 conf.
 : file. i.e. the only thing that was different on our machines as
 compared to
 : everyone else and moving that conf file got the tests to fail for me. It
 : now fails fairly consistently without the patch (from SOLR-7468) and has
 : been looking good with the patch.

 that smells like the kind of thing that sould have some assume sanity
 checks built into it.

 Given:
 * the test setups a special env before running the test
 * the test assumes the specific env will exist
 * the user's machine may already have env properties set before running
 ant that affect the expected special env

 therefore: before the test does the setup of the special env, it should
 sanity check that the users basic env doesn't have any properties that
 violate the basic situation.

 so, hypothetical example based on what little i understand the
 authentication stuff: if the purpose of a test is to prove that some
 requests work with (or w/o) kerberos authentication, then before doing any
 setup of a mock kerberos env (or before setting up a mock situation
 where no authentication is required), the test should verify that there
 isn't already an existing kerberos env. (or if possible: unset whatever
 env/properties define that env)


 trivial example of a similar situation is the script engine tests --
 TestBadConfigs.testBogusScriptEngine:  the purpose of the test is to
 ensure that a solrconfig.xml file that refers to a script engine (by
 name) which is not installed on the machine will produce an expeted error
 at Solr init.  before doing the Solr init, we have an whitebox assume that
 asks the JVM directly if a script engine with that name already exists)



 -Hoss
 http://www.lucidworks.com/

 -
 To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
 For additional commands, e-mail: dev-h...@lucene.apache.org




 --
 Anshum Gupta



[JENKINS] Lucene-Solr-trunk-Linux (64bit/jdk1.8.0_45) - Build # 12841 - Failure!

2015-05-27 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-Linux/12841/
Java: 64bit/jdk1.8.0_45 -XX:-UseCompressedOops -XX:+UseParallelGC

3 tests failed.
FAILED:  
junit.framework.TestSuite.org.apache.solr.handler.TestSolrConfigHandlerCloud

Error Message:
ERROR: SolrIndexSearcher opens=281 closes=280

Stack Trace:
java.lang.AssertionError: ERROR: SolrIndexSearcher opens=281 closes=280
at __randomizedtesting.SeedInfo.seed([CE5375984EAAF35C]:0)
at org.junit.Assert.fail(Assert.java:93)
at 
org.apache.solr.SolrTestCaseJ4.endTrackingSearchers(SolrTestCaseJ4.java:497)
at org.apache.solr.SolrTestCaseJ4.afterClass(SolrTestCaseJ4.java:232)
at sun.reflect.GeneratedMethodAccessor41.invoke(Unknown Source)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:799)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at java.lang.Thread.run(Thread.java:745)


FAILED:  
junit.framework.TestSuite.org.apache.solr.handler.TestSolrConfigHandlerCloud

Error Message:
2 threads leaked from SUITE scope at 
org.apache.solr.handler.TestSolrConfigHandlerCloud: 1) Thread[id=2806, 
name=searcherExecutor-1604-thread-1, state=WAITING, 
group=TGRP-TestSolrConfigHandlerCloud] at sun.misc.Unsafe.park(Native 
Method) at 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
 at 
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) 
at 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1067)   
  at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1127) 
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) 
at java.lang.Thread.run(Thread.java:745)2) Thread[id=2097, 
name=qtp239723857-2097, state=RUNNABLE, group=TGRP-TestSolrConfigHandlerCloud]  
   at java.util.WeakHashMap.get(WeakHashMap.java:403) at 
org.apache.solr.servlet.cache.HttpCacheHeaderUtil.calcEtag(HttpCacheHeaderUtil.java:101)
 at 
org.apache.solr.servlet.cache.HttpCacheHeaderUtil.doCacheHeaderValidation(HttpCacheHeaderUtil.java:219)
 at org.apache.solr.servlet.HttpSolrCall.call(HttpSolrCall.java:428)
 at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:227)
 at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:196)
 at 
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652)
 at 
org.apache.solr.client.solrj.embedded.JettySolrRunner$DebugFilter.doFilter(JettySolrRunner.java:105)
 at 
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1652)
 at 
org.eclipse.jetty.servlets.UserAgentFilter.doFilter(UserAgentFilter.java:83)
 at 

[jira] [Commented] (LUCENE-6371) Improve Spans payload collection

2015-05-27 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6371?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560786#comment-14560786
 ] 

Robert Muir commented on LUCENE-6371:
-

But that does not explain the goal. I want to know the reasoning, what is the 
tradeoff, what are we gaining by having LUCENE-6466 changes? Thats the big 
thing I am missing.

Are there performance improvements in the benchmark? Otherwise I would rather 
not have stuff like SpanSimilarity or null TermContexts maps in ctors, unless 
its really needed for some reason.

 Improve Spans payload collection
 

 Key: LUCENE-6371
 URL: https://issues.apache.org/jira/browse/LUCENE-6371
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Paul Elschot
Assignee: Alan Woodward
Priority: Minor
 Fix For: Trunk, 5.3

 Attachments: LUCENE-6371.patch, LUCENE-6371.patch, LUCENE-6371.patch, 
 LUCENE-6371.patch


 Spin off from LUCENE-6308, see the comments there from around 23 March 2015.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6371) Improve Spans payload collection

2015-05-27 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6371?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560756#comment-14560756
 ] 

Robert Muir commented on LUCENE-6371:
-

But what does it buy? I don't like having two apis to accomplish the same 
thing. In this case i don't think we really save anything.

 Improve Spans payload collection
 

 Key: LUCENE-6371
 URL: https://issues.apache.org/jira/browse/LUCENE-6371
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Paul Elschot
Assignee: Alan Woodward
Priority: Minor
 Fix For: Trunk, 5.3

 Attachments: LUCENE-6371.patch, LUCENE-6371.patch, LUCENE-6371.patch, 
 LUCENE-6371.patch


 Spin off from LUCENE-6308, see the comments there from around 23 March 2015.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6501) Flatten subreader structure in ParallelCompositeReader

2015-05-27 Thread Uwe Schindler (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6501?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560841#comment-14560841
 ] 

Uwe Schindler commented on LUCENE-6501:
---

No problems with testing up to now. I think this is safe for 5.3 (not 5.2) so 
it can bake longer.

 Flatten subreader structure in ParallelCompositeReader
 --

 Key: LUCENE-6501
 URL: https://issues.apache.org/jira/browse/LUCENE-6501
 Project: Lucene - Core
  Issue Type: Improvement
  Components: core/index
Affects Versions: 5.2
Reporter: Uwe Schindler
Assignee: Uwe Schindler
 Fix For: Trunk, 5.3

 Attachments: LUCENE-6501.patch


 The current implementation of ParallelCompositeReader reassembles the whole 
 subreader structure of the wrapped reader with ParallelLeafReader and 
 ParallelCompositeReader.
 This leads to bugs like described in LUCENE-6500. This reaches back to the 
 time when this reader was reimplemented for the first time shortly before 
 release of 4.0. Shortly afterwards, we completely changed our search 
 infrastructure to just call leaves() and working with them. The method 
 getSequentialSubReaders was made protected, just to be implemented by 
 subclasses (like this one). But no external code can ever call it. Also the 
 search API just rely on the baseId in relation to the top-level reader (to 
 correctly present document ids). The structure is completely unimportant.
 This issue will therefore simplify ParallelCompositeReader to just fetch all 
 LeafReaders and build a flat structure of ParallelLeafReaders from it. This 
 also has the nice side-effect, that only the parallel leaf readers must be 
 equally sized, not their structure.
 This issue will solve LUCENE-6500 as a side effect. I just opened a new issue 
 for discussion and to have this listed as feature and not bug.
 In general, we could also hide the ParallelLeafReader class and make it an 
 implementation detail. ParallelCompositeReader would be the only entry point 
 - because people could pass any IndexReader structure in, a single 
 AtomicReader would just produce a CompositeReader with one leaf. We could 
 then also rename it back to ParallelReader (like it was in pre Lucene4).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



NASA/JPL's involvement in Memex: now public

2015-05-27 Thread Mattmann, Chris A (3980)
Hey Folks,

Just wanted to share publicly some articles recently on NASA and
JPL’s involvement in Memex. It’s basically focused around Tika,
Nutch and Solr, so keep up the great work on all projects. A sampling
of the recent press/articles:

E. Landau. Deep Web Search May Help Scientists. NASA Jet Propulsion
Laboratory. Web
Feature. May 22, 2015.
http://www.jpl.nasa.gov/news/news.php?feature=4595
http://www.nasa.gov/jpl/deep-web-search-may-help-scientists



R. Chirgwin. JPL joins DARPA’s Memex project: Better search and indexing
for space stuff.
The Register - UK. May 
27, 2015.
http://www.theregister.co.uk/2015/05/27/jpl_joins_darpas_memex_project/



J. Lendino. NASA, DARPA collaborating on Deep Web search to analyze
spacecraft data.
ExtremeTech. May 26, 2015.
http://goo.gl/iRAFXC



M. Prigg. Nasa joins US government project to create ’Google for the dark
net’ that could
uncover cyber criminals, paedophiles and drug dealers in the online
underworld. Daily Mail -
UK. May 25, 2015.
http://goo.gl/xMj2hH



J. Wolman. Knee-deep in data. The Positive. May 27, 2015.
http://thepositive.com/knee-deep-in-data/



Scientists will benefit more from Deep Web Searches. GreenAtom. May 26,
2015.
http://goo.gl/UqDJAN



NASA to put science face on scary DARPA project. TRTWorld.com. May 26,
2015.
http://goo.gl/HJPv5O



Kitware Participates in DARPA Memex Program, Developing Software to
Address Complex
Search Problems. PR News, May 23, 2015.
http://www.prweb.com/releases/2015/05/prweb12741017.htm

Any questions let me know.

Cheers,
Chris









++
Chris Mattmann, Ph.D.
Chief Architect
Instrument Software and Science Data Systems Section (398)
NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
Office: 168-519, Mailstop: 168-527
Email: chris.a.mattm...@nasa.gov
WWW:  http://sunset.usc.edu/~mattmann/
++
Adjunct Associate Professor, Computer Science Department
University of Southern California, Los Angeles, CA 90089 USA
++




[jira] [Commented] (LUCENE-6501) Flatten subreader structure in ParallelCompositeReader

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6501?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560886#comment-14560886
 ] 

ASF subversion and git services commented on LUCENE-6501:
-

Commit 1682000 from [~thetaphi] in branch 'dev/branches/branch_5x'
[ https://svn.apache.org/r1682000 ]

Merged revision(s) 1681998 from lucene/dev/trunk:
LUCENE-6501: Flatten subreader structure in ParallelCompositeReader (fixes 
close listener bug LUCENE-6500)

 Flatten subreader structure in ParallelCompositeReader
 --

 Key: LUCENE-6501
 URL: https://issues.apache.org/jira/browse/LUCENE-6501
 Project: Lucene - Core
  Issue Type: Improvement
  Components: core/index
Affects Versions: 5.2
Reporter: Uwe Schindler
Assignee: Uwe Schindler
 Fix For: Trunk, 5.3

 Attachments: LUCENE-6501.patch


 The current implementation of ParallelCompositeReader reassembles the whole 
 subreader structure of the wrapped reader with ParallelLeafReader and 
 ParallelCompositeReader.
 This leads to bugs like described in LUCENE-6500. This reaches back to the 
 time when this reader was reimplemented for the first time shortly before 
 release of 4.0. Shortly afterwards, we completely changed our search 
 infrastructure to just call leaves() and working with them. The method 
 getSequentialSubReaders was made protected, just to be implemented by 
 subclasses (like this one). But no external code can ever call it. Also the 
 search API just rely on the baseId in relation to the top-level reader (to 
 correctly present document ids). The structure is completely unimportant.
 This issue will therefore simplify ParallelCompositeReader to just fetch all 
 LeafReaders and build a flat structure of ParallelLeafReaders from it. This 
 also has the nice side-effect, that only the parallel leaf readers must be 
 equally sized, not their structure.
 This issue will solve LUCENE-6500 as a side effect. I just opened a new issue 
 for discussion and to have this listed as feature and not bug.
 In general, we could also hide the ParallelLeafReader class and make it an 
 implementation detail. ParallelCompositeReader would be the only entry point 
 - because people could pass any IndexReader structure in, a single 
 AtomicReader would just produce a CompositeReader with one leaf. We could 
 then also rename it back to ParallelReader (like it was in pre Lucene4).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6487) Add WGS84 capability to geo3d support

2015-05-27 Thread David Smiley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6487?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560979#comment-14560979
 ] 

David Smiley commented on LUCENE-6487:
--

This looks ready to me.  I do have a couple lingering thoughts (_not blockers_):
* Might it make sense for Vector to to have it's x, y, and z, plus GeoPoint's 
magnitude as floats instead of doubles?   I was just thinking the objects are 
comparatively heavy compared to a lat-lon based Point.
* It'd be nice if there was some random round-trip tests from lat  lon to 
GeoPoint to back again, checking the result is within a tolerance.

[~nknize] do you have an interest in reviewing the Geo3d lucene6487 branch with 
respect to WGS84 support?

 Add WGS84 capability to geo3d support
 -

 Key: LUCENE-6487
 URL: https://issues.apache.org/jira/browse/LUCENE-6487
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/spatial
Reporter: Karl Wright
 Attachments: LUCENE-6487.patch, LUCENE-6487.patch, LUCENE-6487.patch, 
 LUCENE-6487.patch


 WGS84 compatibility has been requested for geo3d.  This involves working with 
 an ellipsoid rather than a unit sphere.  The general formula for an ellipsoid 
 is:
 x^2/a^2 + y^2/b^2 + z^2/c^2 = 1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6501) Flatten subreader structure in ParallelCompositeReader

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6501?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560875#comment-14560875
 ] 

ASF subversion and git services commented on LUCENE-6501:
-

Commit 1681998 from [~thetaphi] in branch 'dev/trunk'
[ https://svn.apache.org/r1681998 ]

LUCENE-6501: Flatten subreader structure in ParallelCompositeReader (fixes 
close listener bug LUCENE-6500)

 Flatten subreader structure in ParallelCompositeReader
 --

 Key: LUCENE-6501
 URL: https://issues.apache.org/jira/browse/LUCENE-6501
 Project: Lucene - Core
  Issue Type: Improvement
  Components: core/index
Affects Versions: 5.2
Reporter: Uwe Schindler
Assignee: Uwe Schindler
 Fix For: Trunk, 5.3

 Attachments: LUCENE-6501.patch


 The current implementation of ParallelCompositeReader reassembles the whole 
 subreader structure of the wrapped reader with ParallelLeafReader and 
 ParallelCompositeReader.
 This leads to bugs like described in LUCENE-6500. This reaches back to the 
 time when this reader was reimplemented for the first time shortly before 
 release of 4.0. Shortly afterwards, we completely changed our search 
 infrastructure to just call leaves() and working with them. The method 
 getSequentialSubReaders was made protected, just to be implemented by 
 subclasses (like this one). But no external code can ever call it. Also the 
 search API just rely on the baseId in relation to the top-level reader (to 
 correctly present document ids). The structure is completely unimportant.
 This issue will therefore simplify ParallelCompositeReader to just fetch all 
 LeafReaders and build a flat structure of ParallelLeafReaders from it. This 
 also has the nice side-effect, that only the parallel leaf readers must be 
 equally sized, not their structure.
 This issue will solve LUCENE-6500 as a side effect. I just opened a new issue 
 for discussion and to have this listed as feature and not bug.
 In general, we could also hide the ParallelLeafReader class and make it an 
 implementation detail. ParallelCompositeReader would be the only entry point 
 - because people could pass any IndexReader structure in, a single 
 AtomicReader would just produce a CompositeReader with one leaf. We could 
 then also rename it back to ParallelReader (like it was in pre Lucene4).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6500) ParallelCompositeReader does not always call closed listeners

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6500?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560878#comment-14560878
 ] 

ASF subversion and git services commented on LUCENE-6500:
-

Commit 1681998 from [~thetaphi] in branch 'dev/trunk'
[ https://svn.apache.org/r1681998 ]

LUCENE-6501: Flatten subreader structure in ParallelCompositeReader (fixes 
close listener bug LUCENE-6500)

 ParallelCompositeReader does not always call closed listeners
 -

 Key: LUCENE-6500
 URL: https://issues.apache.org/jira/browse/LUCENE-6500
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Adrien Grand
Assignee: Adrien Grand
Priority: Minor
 Attachments: LUCENE-6500.patch


 CompositeParallelReader misses to call closed listeners when the reader which 
 is provided at construction time does not wrap leaf readers directly, such as 
 a multi reader over directory readers.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: NASA/JPL's involvement in Memex: now public

2015-05-27 Thread Karl Wright
The name Memex is, I think, already in use by a company that was bought
out by Autonomy some years ago, and thence by HP.  I presume this is *not*
what you are talking about?

Thanks,
Karl


On Wed, May 27, 2015 at 8:26 AM, Mattmann, Chris A (3980) 
chris.a.mattm...@jpl.nasa.gov wrote:

 Hey Folks,

 Just wanted to share publicly some articles recently on NASA and
 JPL’s involvement in Memex. It’s basically focused around Tika,
 Nutch and Solr, so keep up the great work on all projects. A sampling
 of the recent press/articles:

 E. Landau. Deep Web Search May Help Scientists. NASA Jet Propulsion
 Laboratory. Web
 Feature. May 22, 2015.
 http://www.jpl.nasa.gov/news/news.php?feature=4595
 http://www.nasa.gov/jpl/deep-web-search-may-help-scientists



 R. Chirgwin. JPL joins DARPA’s Memex project: Better search and indexing
 for space stuff.
 The Register - UK.
 May 27, 2015.
 http://www.theregister.co.uk/2015/05/27/jpl_joins_darpas_memex_project/



 J. Lendino. NASA, DARPA collaborating on Deep Web search to analyze
 spacecraft data.
 ExtremeTech. May 26, 2015.
 http://goo.gl/iRAFXC



 M. Prigg. Nasa joins US government project to create ’Google for the dark
 net’ that could
 uncover cyber criminals, paedophiles and drug dealers in the online
 underworld. Daily Mail -
 UK. May 25, 2015.
 http://goo.gl/xMj2hH



 J. Wolman. Knee-deep in data. The Positive. May 27, 2015.
 http://thepositive.com/knee-deep-in-data/



 Scientists will benefit more from Deep Web Searches. GreenAtom. May 26,
 2015.
 http://goo.gl/UqDJAN



 NASA to put science face on scary DARPA project. TRTWorld.com. May 26,
 2015.
 http://goo.gl/HJPv5O



 Kitware Participates in DARPA Memex Program, Developing Software to
 Address Complex
 Search Problems. PR News, May 23, 2015.
 http://www.prweb.com/releases/2015/05/prweb12741017.htm

 Any questions let me know.

 Cheers,
 Chris









 ++
 Chris Mattmann, Ph.D.
 Chief Architect
 Instrument Software and Science Data Systems Section (398)
 NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
 Office: 168-519, Mailstop: 168-527
 Email: chris.a.mattm...@nasa.gov
 WWW:  http://sunset.usc.edu/~mattmann/
 ++
 Adjunct Associate Professor, Computer Science Department
 University of Southern California, Los Angeles, CA 90089 USA
 ++





[jira] [Commented] (SOLR-7146) MiniSolrCloudCluster based tests can fail with ZooKeeperException NoNode for /live_nodes

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7146?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560903#comment-14560903
 ] 

ASF subversion and git services commented on SOLR-7146:
---

Commit 1682002 from sha...@apache.org in branch 'dev/trunk'
[ https://svn.apache.org/r1682002 ]

SOLR-7146: MiniSolrCloudCluster based tests can fail with ZooKeeperException 
NoNode for /live_nodes

 MiniSolrCloudCluster based tests can fail with ZooKeeperException NoNode for 
 /live_nodes
 

 Key: SOLR-7146
 URL: https://issues.apache.org/jira/browse/SOLR-7146
 Project: Solr
  Issue Type: Bug
  Components: SolrCloud, Tests
Reporter: Shalin Shekhar Mangar
Assignee: Shalin Shekhar Mangar
Priority: Minor
 Fix For: Trunk, 5.2

 Attachments: SOLR-7146.patch, SOLR-7146v2.patch


 MiniSolrCloudCluster based tests can fail with the following exception:
 {code}
 org.apache.solr.common.cloud.ZooKeeperException: 
   at 
 __randomizedtesting.SeedInfo.seed([3F3D838A8ADC9385:F153ADFBF163EC6D]:0)
   at 
 org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:463)
   at 
 org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:763)
   at 
 org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:752)
   at 
 org.apache.solr.cloud.MiniSolrCloudCluster.createCollection(MiniSolrCloudCluster.java:193)
   at 
 org.apache.solr.handler.component.TestTwoPhaseDistributedQuery.testNoExtraFieldsRequestedFromShardsInPhaseOne(TestTwoPhaseDistributedQuery.java:79)
   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877)
   at 
 com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
   at 
 org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
   at 
 org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
   at 
 com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
   at 
 org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
   at 
 org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
   at 
 org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
   at 
 com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   at 
 com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
   at 
 com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
   at 
 com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783)
   at 
 com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   at 
 com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
   at 
 org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
   at 
 org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
   at 
 com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
   at 
 com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
   at 
 

Re: [JENKINS] Lucene-Solr-5.x-Linux (64bit/jdk1.8.0_45) - Build # 12644 - Failure!

2015-05-27 Thread Mark Miller
Oh baby, I need a new computer soon...

On Wed, May 27, 2015 at 5:08 AM Anshum Gupta ans...@anshumgupta.net wrote:

 8-real-core Haswell-E with 64G DDR4 memory and a NVMe 750-series SSD.

 Can run *all* of the Lucene and Solr tests in 10 minutes by running
 multiple ant jobs in parallel!

 On Wed, May 27, 2015 at 1:17 AM, Ramkumar R. Aiyengar 
 andyetitmo...@gmail.com wrote:

 Curious.. sarowe, what's the spec?
 On 26 May 2015 20:41, Anshum Gupta ans...@anshumgupta.net wrote:

 The last buch of fixes seems to have fixed this. The tests passed on a
 Jenkins that had failing tests earlier.
 Thanks Steve Rowe for lending the super-powerful machine that runs the
 entire suite in 8 min!

 I've seen about 20 runs on that box and also runs on Policeman Jenkins
 with no issues related to this test since the last commit so I've also
 back-ported this to 5x as well.

 On Tue, May 26, 2015 at 9:19 AM, Chris Hostetter 
 hossman_luc...@fucit.org wrote:


 : Right, as I said, we weren't hitting this issue due to our Kerberos
 conf.
 : file. i.e. the only thing that was different on our machines as
 compared to
 : everyone else and moving that conf file got the tests to fail for me.
 It
 : now fails fairly consistently without the patch (from SOLR-7468) and
 has
 : been looking good with the patch.

 that smells like the kind of thing that sould have some assume sanity
 checks built into it.

 Given:
 * the test setups a special env before running the test
 * the test assumes the specific env will exist
 * the user's machine may already have env properties set before running
 ant that affect the expected special env

 therefore: before the test does the setup of the special env, it should
 sanity check that the users basic env doesn't have any properties that
 violate the basic situation.

 so, hypothetical example based on what little i understand the
 authentication stuff: if the purpose of a test is to prove that some
 requests work with (or w/o) kerberos authentication, then before doing
 any
 setup of a mock kerberos env (or before setting up a mock situation
 where no authentication is required), the test should verify that there
 isn't already an existing kerberos env. (or if possible: unset
 whatever
 env/properties define that env)


 trivial example of a similar situation is the script engine tests --
 TestBadConfigs.testBogusScriptEngine:  the purpose of the test is to
 ensure that a solrconfig.xml file that refers to a script engine (by
 name) which is not installed on the machine will produce an expeted
 error
 at Solr init.  before doing the Solr init, we have an whitebox assume
 that
 asks the JVM directly if a script engine with that name already exists)



 -Hoss
 http://www.lucidworks.com/

 -
 To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
 For additional commands, e-mail: dev-h...@lucene.apache.org




 --
 Anshum Gupta




 --
 Anshum Gupta



[jira] [Created] (LUCENE-6504) implement norms with random access API

2015-05-27 Thread Robert Muir (JIRA)
Robert Muir created LUCENE-6504:
---

 Summary: implement norms with random access API
 Key: LUCENE-6504
 URL: https://issues.apache.org/jira/browse/LUCENE-6504
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Robert Muir


We added this api in LUCENE-5729 but we never explored implementing norms with 
it. These are generally the largest consumer of heap memory and often a real 
hassle for users.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-7599) Remove cruft from SolrCloud tests

2015-05-27 Thread Shalin Shekhar Mangar (JIRA)
Shalin Shekhar Mangar created SOLR-7599:
---

 Summary: Remove cruft from SolrCloud tests
 Key: SOLR-7599
 URL: https://issues.apache.org/jira/browse/SOLR-7599
 Project: Solr
  Issue Type: Task
  Components: SolrCloud, Tests
Reporter: Shalin Shekhar Mangar
Assignee: Shalin Shekhar Mangar
 Fix For: Trunk, 5.3


I see many tests which blindly have distribSetUp and distribTearDown methods 
setting a variety of options and system properties that aren't required 
anymore. 

This is because some base test classes have been refactored such that these 
options are redundant. In other cases, people have copied the structure of 
tests blindly instead of understanding what each parameter does.

Let's try to remove the unnecessary config params from such tests.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-5.x-Linux (64bit/jdk1.8.0_45) - Build # 12644 - Failure!

2015-05-27 Thread Steve Rowe
SSD: http://www.newegg.com/Product/Product.aspx?Item=N82E16820167299
CPU: http://www.newegg.com/Product/Product.aspx?Item=N82E16819117404
M/B: http://www.newegg.com/Product/Product.aspx?Item=N82E16813132518
RAM: http://www.newegg.com/Product/Product.aspx?Item=N82E16820231820

The mem wasn’t listed as supported by the mobo manufacturer, and it isn’t 
detected at its full speed (2800MHz), so currently running at 2400 
(“overclocked” from detected 2100 I think).  CPU isn’t overclocked from stock 
3GHz, but I got a liquid cooler, thinking I’d experiment (haven’t much yet).  
Even without overclocking the fans spin faster when all the cores are busy, and 
it’s quite the little space heater.

I installed Debian 8, but had to fix the installer in a couple places because 
it didn’t know about the new NVMe device naming scheme:

https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=785147
https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=785149

I also upgraded to the 4.0 Linux kernel, since Debian 8 ships with the 3.16 
kernel, and 3.19 contains a bunch of NVMe improvements.

And I turned “swappiness down to zero (thanks to Mike: 
http://blog.mikemccandless.com/2011/04/just-say-no-to-swapping.html) after 
seeing a bunch of test stalls while running the Lucene monster tests with 4 
JVMs (takes about 2 hours, but I can get it down to 90 minutes or so by 
splitting the two tests in Test2BSortedDocValues out into their own suites - 
I’ll make an issue).

Steve

 On May 27, 2015, at 5:08 AM, Anshum Gupta ans...@anshumgupta.net wrote:
 
 8-real-core Haswell-E with 64G DDR4 memory and a NVMe 750-series SSD.
 
 Can run *all* of the Lucene and Solr tests in 10 minutes by running multiple 
 ant jobs in parallel!
 
 On Wed, May 27, 2015 at 1:17 AM, Ramkumar R. Aiyengar 
 andyetitmo...@gmail.com wrote:
 Curious.. sarowe, what's the spec?
 
 On 26 May 2015 20:41, Anshum Gupta ans...@anshumgupta.net wrote:
 The last buch of fixes seems to have fixed this. The tests passed on a 
 Jenkins that had failing tests earlier.
 Thanks Steve Rowe for lending the super-powerful machine that runs the entire 
 suite in 8 min! 
 
 I've seen about 20 runs on that box and also runs on Policeman Jenkins with 
 no issues related to this test since the last commit so I've also back-ported 
 this to 5x as well.
 
 On Tue, May 26, 2015 at 9:19 AM, Chris Hostetter hossman_luc...@fucit.org 
 wrote:
 
 : Right, as I said, we weren't hitting this issue due to our Kerberos conf.
 : file. i.e. the only thing that was different on our machines as compared to
 : everyone else and moving that conf file got the tests to fail for me. It
 : now fails fairly consistently without the patch (from SOLR-7468) and has
 : been looking good with the patch.
 
 that smells like the kind of thing that sould have some assume sanity
 checks built into it.
 
 Given:
 * the test setups a special env before running the test
 * the test assumes the specific env will exist
 * the user's machine may already have env properties set before running ant 
 that affect the expected special env
 
 therefore: before the test does the setup of the special env, it should
 sanity check that the users basic env doesn't have any properties that
 violate the basic situation.
 
 so, hypothetical example based on what little i understand the
 authentication stuff: if the purpose of a test is to prove that some
 requests work with (or w/o) kerberos authentication, then before doing any
 setup of a mock kerberos env (or before setting up a mock situation
 where no authentication is required), the test should verify that there
 isn't already an existing kerberos env. (or if possible: unset whatever
 env/properties define that env)
 
 
 trivial example of a similar situation is the script engine tests --
 TestBadConfigs.testBogusScriptEngine:  the purpose of the test is to
 ensure that a solrconfig.xml file that refers to a script engine (by
 name) which is not installed on the machine will produce an expeted error
 at Solr init.  before doing the Solr init, we have an whitebox assume that
 asks the JVM directly if a script engine with that name already exists)
 
 
 
 -Hoss
 http://www.lucidworks.com/
 
 -
 To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
 For additional commands, e-mail: dev-h...@lucene.apache.org
 
 
 
 
 -- 
 Anshum Gupta
 
 
 
 -- 
 Anshum Gupta


-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-7590) Finish and improve MDC logging support.

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7590?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561005#comment-14561005
 ] 

ASF subversion and git services commented on SOLR-7590:
---

Commit 1682031 from [~markrmil...@gmail.com] in branch 'dev/branches/branch_5x'
[ https://svn.apache.org/r1682031 ]

SOLR-7590: Finish and improve MDC context logging support.

 Finish and improve MDC logging support.
 ---

 Key: SOLR-7590
 URL: https://issues.apache.org/jira/browse/SOLR-7590
 Project: Solr
  Issue Type: Improvement
Reporter: Mark Miller
Assignee: Mark Miller
 Fix For: Trunk, 5.3

 Attachments: SOLR-7950.patch, SOLR-7950.patch, SOLR-7950.patch, 
 SOLR-7950.patch, SOLR-7950.patch, SOLR-7950.patch, SOLR-7950.patch, 
 SOLR-7950.patch, SOLR-7950.patch, SOLR-7950.patch, SOLR-7950.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4506) [solr4.0.0] many index.{date} dir in replication node

2015-05-27 Thread Timothy Potter (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4506?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561026#comment-14561026
 ] 

Timothy Potter commented on SOLR-4506:
--

Hi [~markrmil...@gmail.com], I'd like to take this one up as I need it for some 
other issue I'm working on. Let me know if you have any updated code or ideas 
on this otherwise, feel free to assign over to me and I'll work on it for 5.3. 
Thanks

 [solr4.0.0] many index.{date} dir in replication node 
 --

 Key: SOLR-4506
 URL: https://issues.apache.org/jira/browse/SOLR-4506
 Project: Solr
  Issue Type: Bug
  Components: SolrCloud
Affects Versions: 4.0
 Environment: the solr4.0 runs on suse11.
 mem:32G
 cpu:16 cores
Reporter: zhuojunjian
Assignee: Mark Miller
Priority: Minor
 Fix For: 4.9, Trunk

   Original Estimate: 12h
  Remaining Estimate: 12h

 in our test,we used solrcloud feature in solr4.0(version detail 
 :4.0.0.2012.10.06.03.04.33).
 the solrcloud configuration is 3 shards and 2 replications each shard.
 we found that there are over than 25 dirs which named index.{date} in one 
 replication node belonging to shard 3. 
 for example:
 index.2013021725864  index.20130218012211880  index.20130218015714713  
 index.20130218023101958  index.20130218030424083  tlog
 index.20130218005648324  index.20130218012751078  index.20130218020141293  
 the issue seems like SOLR-1781. but it is fixed in 4.0-BETA,5.0. 
 so is solr4.0 ? if it is fixed too in solr4.0, why we find the issue again ?
 what can I do?   



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6504) implement norms with random access API

2015-05-27 Thread Uwe Schindler (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6504?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561025#comment-14561025
 ] 

Uwe Schindler commented on LUCENE-6504:
---

There are some unrelated changes in ByteBufferIndexInput: You removed the 
offset=0 implementation. Did you find out that this optimization for offset=0 
brings no additional performance?

 implement norms with random access API
 --

 Key: LUCENE-6504
 URL: https://issues.apache.org/jira/browse/LUCENE-6504
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Robert Muir
 Attachments: LUCENE-6504.patch


 We added this api in LUCENE-5729 but we never explored implementing norms 
 with it. These are generally the largest consumer of heap memory and often a 
 real hassle for users.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-7587) TestSpellCheckResponse stalled and never timed out -- possible VersionBucket bug? (5.2 branch)

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7587?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560969#comment-14560969
 ] 

ASF subversion and git services commented on SOLR-7587:
---

Commit 1682019 from [~thelabdude] in branch 'dev/branches/lucene_solr_5_2'
[ https://svn.apache.org/r1682019 ]

SOLR-7587: Move the call to seed version buckets to before buffering updates 
during core construction

 TestSpellCheckResponse stalled and never timed out -- possible VersionBucket 
 bug? (5.2 branch)
 --

 Key: SOLR-7587
 URL: https://issues.apache.org/jira/browse/SOLR-7587
 Project: Solr
  Issue Type: Bug
Reporter: Hoss Man
Assignee: Timothy Potter
Priority: Blocker
 Fix For: Trunk, 5.2

 Attachments: SOLR-7587.patch, SOLR-7587_minor_refactor.patch, 
 jstack.1.txt, jstack.2.txt, junit4-J0-20150522_181244_599.events, 
 junit4-J0-20150522_181244_599.spill, junit4-J0-20150522_181244_599.suites


 On the 5.2 branch (r1681250), I encountered a solrj test stalled for over 110 
 minutes before i finally killed it...
 {noformat}
[junit4] Suite: org.apache.solr.common.util.TestRetryUtil
[junit4] Completed [55/60] on J1 in 1.04s, 1 test
[junit4] 
[junit4] HEARTBEAT J0 PID(12147@tray): 2015-05-22T18:14:56, stalled for  
 121s at: TestSpellCheckResponse.testSpellCheckResponse
[junit4] HEARTBEAT J0 PID(12147@tray): 2015-05-22T18:15:56, stalled for  
 181s at: TestSpellCheckResponse.testSpellCheckResponse
 ...
[junit4] HEARTBEAT J0 PID(12147@tray): 2015-05-22T20:00:56, stalled for 
 6481s at: TestSpellCheckResponse.testSpellCheckResponse
[junit4] HEARTBEAT J0 PID(12147@tray): 2015-05-22T20:01:56, stalled for 
 6541s at: TestSpellCheckResponse.testSpellCheckResponse
[junit4] HEARTBEAT J0 PID(12147@tray): 2015-05-22T20:02:56, stalled for 
 6601s at: TestSpellCheckResponse.testSpellCheckResponse
 {noformat}
 I'll attach some jstack output as well as all the temp files from the J0 
 runner.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-7599) Remove cruft from SolrCloud tests

2015-05-27 Thread Shalin Shekhar Mangar (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7599?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560968#comment-14560968
 ] 

Shalin Shekhar Mangar commented on SOLR-7599:
-

Here's a list of stuff that should be looked at:
# Does a test really need distribSetUp and distribTearDown methods?
# Usefulness of useJettyDataDir
# solr.xml.persist
# The need to change the value for DirectUpdateHandler2.commitOnClose. Only a 
few tests really need it.
# Setting system properties such as System.setProperty(numShards, 
Integer.toString(sliceCount));
# checkCreatedVsState -- looks like it is always false across all tests.

 Remove cruft from SolrCloud tests
 -

 Key: SOLR-7599
 URL: https://issues.apache.org/jira/browse/SOLR-7599
 Project: Solr
  Issue Type: Task
  Components: SolrCloud, Tests
Reporter: Shalin Shekhar Mangar
Assignee: Shalin Shekhar Mangar
 Fix For: Trunk, 5.3


 I see many tests which blindly have distribSetUp and distribTearDown methods 
 setting a variety of options and system properties that aren't required 
 anymore. 
 This is because some base test classes have been refactored such that these 
 options are redundant. In other cases, people have copied the structure of 
 tests blindly instead of understanding what each parameter does.
 Let's try to remove the unnecessary config params from such tests.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-7587) TestSpellCheckResponse stalled and never timed out -- possible VersionBucket bug? (5.2 branch)

2015-05-27 Thread Timothy Potter (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-7587?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Timothy Potter resolved SOLR-7587.
--
   Resolution: Fixed
Fix Version/s: Trunk

 TestSpellCheckResponse stalled and never timed out -- possible VersionBucket 
 bug? (5.2 branch)
 --

 Key: SOLR-7587
 URL: https://issues.apache.org/jira/browse/SOLR-7587
 Project: Solr
  Issue Type: Bug
Reporter: Hoss Man
Assignee: Timothy Potter
Priority: Blocker
 Fix For: Trunk, 5.2

 Attachments: SOLR-7587.patch, SOLR-7587_minor_refactor.patch, 
 jstack.1.txt, jstack.2.txt, junit4-J0-20150522_181244_599.events, 
 junit4-J0-20150522_181244_599.spill, junit4-J0-20150522_181244_599.suites


 On the 5.2 branch (r1681250), I encountered a solrj test stalled for over 110 
 minutes before i finally killed it...
 {noformat}
[junit4] Suite: org.apache.solr.common.util.TestRetryUtil
[junit4] Completed [55/60] on J1 in 1.04s, 1 test
[junit4] 
[junit4] HEARTBEAT J0 PID(12147@tray): 2015-05-22T18:14:56, stalled for  
 121s at: TestSpellCheckResponse.testSpellCheckResponse
[junit4] HEARTBEAT J0 PID(12147@tray): 2015-05-22T18:15:56, stalled for  
 181s at: TestSpellCheckResponse.testSpellCheckResponse
 ...
[junit4] HEARTBEAT J0 PID(12147@tray): 2015-05-22T20:00:56, stalled for 
 6481s at: TestSpellCheckResponse.testSpellCheckResponse
[junit4] HEARTBEAT J0 PID(12147@tray): 2015-05-22T20:01:56, stalled for 
 6541s at: TestSpellCheckResponse.testSpellCheckResponse
[junit4] HEARTBEAT J0 PID(12147@tray): 2015-05-22T20:02:56, stalled for 
 6601s at: TestSpellCheckResponse.testSpellCheckResponse
 {noformat}
 I'll attach some jstack output as well as all the temp files from the J0 
 runner.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-7590) Finish and improve MDC logging support.

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7590?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560970#comment-14560970
 ] 

ASF subversion and git services commented on SOLR-7590:
---

Commit 1682020 from [~markrmil...@gmail.com] in branch 'dev/trunk'
[ https://svn.apache.org/r1682020 ]

SOLR-7590: Finish and improve MDC context logging support.

 Finish and improve MDC logging support.
 ---

 Key: SOLR-7590
 URL: https://issues.apache.org/jira/browse/SOLR-7590
 Project: Solr
  Issue Type: Improvement
Reporter: Mark Miller
Assignee: Mark Miller
 Fix For: Trunk, 5.3

 Attachments: SOLR-7950.patch, SOLR-7950.patch, SOLR-7950.patch, 
 SOLR-7950.patch, SOLR-7950.patch, SOLR-7950.patch, SOLR-7950.patch, 
 SOLR-7950.patch, SOLR-7950.patch, SOLR-7950.patch, SOLR-7950.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6487) Add WGS84 capability to geo3d support

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6487?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560971#comment-14560971
 ] 

ASF subversion and git services commented on LUCENE-6487:
-

Commit 1682021 from [~dsmiley] in branch 'dev/branches/lucene6487'
[ https://svn.apache.org/r1682021 ]

LUCENE-6487: Geo3D with WGS84 patch from Karl: GeoPoint.getLat  getLon.

 Add WGS84 capability to geo3d support
 -

 Key: LUCENE-6487
 URL: https://issues.apache.org/jira/browse/LUCENE-6487
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/spatial
Reporter: Karl Wright
 Attachments: LUCENE-6487.patch, LUCENE-6487.patch, LUCENE-6487.patch, 
 LUCENE-6487.patch


 WGS84 compatibility has been requested for geo3d.  This involves working with 
 an ellipsoid rather than a unit sphere.  The general formula for an ellipsoid 
 is:
 x^2/a^2 + y^2/b^2 + z^2/c^2 = 1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6481) Improve GeoPointField type to only visit high precision boundary terms

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6481?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560985#comment-14560985
 ] 

ASF subversion and git services commented on LUCENE-6481:
-

Commit 1682029 from [~mikemccand] in branch 'dev/branches/LUCENE-6481'
[ https://svn.apache.org/r1682029 ]

LUCENE-6481: validate lat/lon passed to the queries

 Improve GeoPointField type to only visit high precision boundary terms 
 ---

 Key: LUCENE-6481
 URL: https://issues.apache.org/jira/browse/LUCENE-6481
 Project: Lucene - Core
  Issue Type: Improvement
  Components: core/index
Reporter: Nicholas Knize
 Attachments: LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481.patch, 
 LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481_WIP.patch


 Current GeoPointField [LUCENE-6450 | 
 https://issues.apache.org/jira/browse/LUCENE-6450] computes a set of ranges 
 along the space-filling curve that represent a provided bounding box.  This 
 determines which terms to visit in the terms dictionary and which to skip. 
 This is suboptimal for large bounding boxes as we may end up visiting all 
 terms (which could be quite large). 
 This incremental improvement is to improve GeoPointField to only visit high 
 precision terms in boundary ranges and use the postings list for ranges that 
 are completely within the target bounding box.
 A separate improvement is to switch over to auto-prefix and build an 
 Automaton representing the bounding box.  That can be tracked in a separate 
 issue.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-trunk-Linux (32bit/jdk1.8.0_45) - Build # 12843 - Failure!

2015-05-27 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-Linux/12843/
Java: 32bit/jdk1.8.0_45 -client -XX:+UseConcMarkSweepGC

1 tests failed.
FAILED:  
org.apache.solr.cloud.CollectionsAPIAsyncDistributedZkTest.testSolrJAPICalls

Error Message:
Shard split did not complete. Last recorded state: running 
expected:[completed] but was:[running]

Stack Trace:
org.junit.ComparisonFailure: Shard split did not complete. Last recorded state: 
running expected:[completed] but was:[running]
at 
__randomizedtesting.SeedInfo.seed([E204DA36B31551FE:BA605657B57FF92A]:0)
at org.junit.Assert.assertEquals(Assert.java:125)
at 
org.apache.solr.cloud.CollectionsAPIAsyncDistributedZkTest.testSolrJAPICalls(CollectionsAPIAsyncDistributedZkTest.java:101)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:836)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:872)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:886)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:960)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:935)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:845)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:747)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:781)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 

[jira] [Commented] (LUCENE-6504) implement norms with random access API

2015-05-27 Thread Michael McCandless (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6504?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560951#comment-14560951
 ] 

Michael McCandless commented on LUCENE-6504:


+1

 implement norms with random access API
 --

 Key: LUCENE-6504
 URL: https://issues.apache.org/jira/browse/LUCENE-6504
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Robert Muir
 Attachments: LUCENE-6504.patch


 We added this api in LUCENE-5729 but we never explored implementing norms 
 with it. These are generally the largest consumer of heap memory and often a 
 real hassle for users.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-7587) TestSpellCheckResponse stalled and never timed out -- possible VersionBucket bug? (5.2 branch)

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7587?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560964#comment-14560964
 ] 

ASF subversion and git services commented on SOLR-7587:
---

Commit 1682017 from [~thelabdude] in branch 'dev/branches/branch_5x'
[ https://svn.apache.org/r1682017 ]

SOLR-7587: Move the call to seed version buckets to before buffering updates 
during core construction

 TestSpellCheckResponse stalled and never timed out -- possible VersionBucket 
 bug? (5.2 branch)
 --

 Key: SOLR-7587
 URL: https://issues.apache.org/jira/browse/SOLR-7587
 Project: Solr
  Issue Type: Bug
Reporter: Hoss Man
Assignee: Timothy Potter
Priority: Blocker
 Fix For: 5.2

 Attachments: SOLR-7587.patch, SOLR-7587_minor_refactor.patch, 
 jstack.1.txt, jstack.2.txt, junit4-J0-20150522_181244_599.events, 
 junit4-J0-20150522_181244_599.spill, junit4-J0-20150522_181244_599.suites


 On the 5.2 branch (r1681250), I encountered a solrj test stalled for over 110 
 minutes before i finally killed it...
 {noformat}
[junit4] Suite: org.apache.solr.common.util.TestRetryUtil
[junit4] Completed [55/60] on J1 in 1.04s, 1 test
[junit4] 
[junit4] HEARTBEAT J0 PID(12147@tray): 2015-05-22T18:14:56, stalled for  
 121s at: TestSpellCheckResponse.testSpellCheckResponse
[junit4] HEARTBEAT J0 PID(12147@tray): 2015-05-22T18:15:56, stalled for  
 181s at: TestSpellCheckResponse.testSpellCheckResponse
 ...
[junit4] HEARTBEAT J0 PID(12147@tray): 2015-05-22T20:00:56, stalled for 
 6481s at: TestSpellCheckResponse.testSpellCheckResponse
[junit4] HEARTBEAT J0 PID(12147@tray): 2015-05-22T20:01:56, stalled for 
 6541s at: TestSpellCheckResponse.testSpellCheckResponse
[junit4] HEARTBEAT J0 PID(12147@tray): 2015-05-22T20:02:56, stalled for 
 6601s at: TestSpellCheckResponse.testSpellCheckResponse
 {noformat}
 I'll attach some jstack output as well as all the temp files from the J0 
 runner.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6371) Improve Spans payload collection

2015-05-27 Thread Alan Woodward (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6371?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560857#comment-14560857
 ] 

Alan Woodward commented on LUCENE-6371:
---

It's the same logic as https://issues.apache.org/jira/browse/LUCENE-6425 
really.  Spans should only be pulled from a SpanQuery after it has been 
rewritten against a searcher, so it makes sense that they should be on 
SpanWeight rather than SpanQuery.  And it simplifies pulling Spans for use in 
things like highlighting (or payload collection, as here) - before you had to 
rewrite, extract terms, build the termcontexts map and then call getSpans(); 
now you just call IndexSearcher.getNormalizedWeight(), cast to a SpanWeight and 
call getSpans.

 Improve Spans payload collection
 

 Key: LUCENE-6371
 URL: https://issues.apache.org/jira/browse/LUCENE-6371
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Paul Elschot
Assignee: Alan Woodward
Priority: Minor
 Fix For: Trunk, 5.3

 Attachments: LUCENE-6371.patch, LUCENE-6371.patch, LUCENE-6371.patch, 
 LUCENE-6371.patch


 Spin off from LUCENE-6308, see the comments there from around 23 March 2015.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-6500) ParallelCompositeReader does not always call closed listeners

2015-05-27 Thread Uwe Schindler (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-6500?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Uwe Schindler resolved LUCENE-6500.
---
   Resolution: Fixed
Fix Version/s: 5.3
   Trunk
 Assignee: Uwe Schindler  (was: Adrien Grand)

Fixed through subreader flattening in LUCENE-6501. Thanks Adrien for reporting!

 ParallelCompositeReader does not always call closed listeners
 -

 Key: LUCENE-6500
 URL: https://issues.apache.org/jira/browse/LUCENE-6500
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Adrien Grand
Assignee: Uwe Schindler
Priority: Minor
 Fix For: Trunk, 5.3

 Attachments: LUCENE-6500.patch


 CompositeParallelReader misses to call closed listeners when the reader which 
 is provided at construction time does not wrap leaf readers directly, such as 
 a multi reader over directory readers.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-7587) TestSpellCheckResponse stalled and never timed out -- possible VersionBucket bug? (5.2 branch)

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7587?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560961#comment-14560961
 ] 

ASF subversion and git services commented on SOLR-7587:
---

Commit 1682016 from [~thelabdude] in branch 'dev/trunk'
[ https://svn.apache.org/r1682016 ]

SOLR-7587: Move the call to seed version buckets to before buffering updates 
during core construction

 TestSpellCheckResponse stalled and never timed out -- possible VersionBucket 
 bug? (5.2 branch)
 --

 Key: SOLR-7587
 URL: https://issues.apache.org/jira/browse/SOLR-7587
 Project: Solr
  Issue Type: Bug
Reporter: Hoss Man
Assignee: Timothy Potter
Priority: Blocker
 Fix For: 5.2

 Attachments: SOLR-7587.patch, SOLR-7587_minor_refactor.patch, 
 jstack.1.txt, jstack.2.txt, junit4-J0-20150522_181244_599.events, 
 junit4-J0-20150522_181244_599.spill, junit4-J0-20150522_181244_599.suites


 On the 5.2 branch (r1681250), I encountered a solrj test stalled for over 110 
 minutes before i finally killed it...
 {noformat}
[junit4] Suite: org.apache.solr.common.util.TestRetryUtil
[junit4] Completed [55/60] on J1 in 1.04s, 1 test
[junit4] 
[junit4] HEARTBEAT J0 PID(12147@tray): 2015-05-22T18:14:56, stalled for  
 121s at: TestSpellCheckResponse.testSpellCheckResponse
[junit4] HEARTBEAT J0 PID(12147@tray): 2015-05-22T18:15:56, stalled for  
 181s at: TestSpellCheckResponse.testSpellCheckResponse
 ...
[junit4] HEARTBEAT J0 PID(12147@tray): 2015-05-22T20:00:56, stalled for 
 6481s at: TestSpellCheckResponse.testSpellCheckResponse
[junit4] HEARTBEAT J0 PID(12147@tray): 2015-05-22T20:01:56, stalled for 
 6541s at: TestSpellCheckResponse.testSpellCheckResponse
[junit4] HEARTBEAT J0 PID(12147@tray): 2015-05-22T20:02:56, stalled for 
 6601s at: TestSpellCheckResponse.testSpellCheckResponse
 {noformat}
 I'll attach some jstack output as well as all the temp files from the J0 
 runner.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6504) implement norms with random access API

2015-05-27 Thread David Smiley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6504?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560987#comment-14560987
 ] 

David Smiley commented on LUCENE-6504:
--

Nice. Before/now, AFAIK norms take 1 byte per field per doc of heap.  I looked 
over the patch briefly; does this essentially put norms off-heap?  And I 
noticed it has more fidelity than a single byte; up to 8 in fact.  Does this 
patch also bring in accurate norms or is something else required to enable 
actually utilize that?

 implement norms with random access API
 --

 Key: LUCENE-6504
 URL: https://issues.apache.org/jira/browse/LUCENE-6504
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Robert Muir
 Attachments: LUCENE-6504.patch


 We added this api in LUCENE-5729 but we never explored implementing norms 
 with it. These are generally the largest consumer of heap memory and often a 
 real hassle for users.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-6501) Flatten subreader structure in ParallelCompositeReader

2015-05-27 Thread Uwe Schindler (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-6501?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Uwe Schindler resolved LUCENE-6501.
---
Resolution: Fixed

 Flatten subreader structure in ParallelCompositeReader
 --

 Key: LUCENE-6501
 URL: https://issues.apache.org/jira/browse/LUCENE-6501
 Project: Lucene - Core
  Issue Type: Improvement
  Components: core/index
Affects Versions: 5.2
Reporter: Uwe Schindler
Assignee: Uwe Schindler
 Fix For: Trunk, 5.3

 Attachments: LUCENE-6501.patch


 The current implementation of ParallelCompositeReader reassembles the whole 
 subreader structure of the wrapped reader with ParallelLeafReader and 
 ParallelCompositeReader.
 This leads to bugs like described in LUCENE-6500. This reaches back to the 
 time when this reader was reimplemented for the first time shortly before 
 release of 4.0. Shortly afterwards, we completely changed our search 
 infrastructure to just call leaves() and working with them. The method 
 getSequentialSubReaders was made protected, just to be implemented by 
 subclasses (like this one). But no external code can ever call it. Also the 
 search API just rely on the baseId in relation to the top-level reader (to 
 correctly present document ids). The structure is completely unimportant.
 This issue will therefore simplify ParallelCompositeReader to just fetch all 
 LeafReaders and build a flat structure of ParallelLeafReaders from it. This 
 also has the nice side-effect, that only the parallel leaf readers must be 
 equally sized, not their structure.
 This issue will solve LUCENE-6500 as a side effect. I just opened a new issue 
 for discussion and to have this listed as feature and not bug.
 In general, we could also hide the ParallelLeafReader class and make it an 
 implementation detail. ParallelCompositeReader would be the only entry point 
 - because people could pass any IndexReader structure in, a single 
 AtomicReader would just produce a CompositeReader with one leaf. We could 
 then also rename it back to ParallelReader (like it was in pre Lucene4).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6500) ParallelCompositeReader does not always call closed listeners

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6500?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560887#comment-14560887
 ] 

ASF subversion and git services commented on LUCENE-6500:
-

Commit 1682000 from [~thetaphi] in branch 'dev/branches/branch_5x'
[ https://svn.apache.org/r1682000 ]

Merged revision(s) 1681998 from lucene/dev/trunk:
LUCENE-6501: Flatten subreader structure in ParallelCompositeReader (fixes 
close listener bug LUCENE-6500)

 ParallelCompositeReader does not always call closed listeners
 -

 Key: LUCENE-6500
 URL: https://issues.apache.org/jira/browse/LUCENE-6500
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Adrien Grand
Assignee: Adrien Grand
Priority: Minor
 Attachments: LUCENE-6500.patch


 CompositeParallelReader misses to call closed listeners when the reader which 
 is provided at construction time does not wrap leaf readers directly, such as 
 a multi reader over directory readers.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: NASA/JPL's involvement in Memex: now public

2015-05-27 Thread Mattmann, Chris A (3980)
Hi Karl,

Memex is a DARPA initiative to build a next generation search
engine capability for the Deep/Dark web. It’s of course a play
on the famous Vannevar Bush paper in 1945:

http://www.darpa.mil/newsevents/releases/2014/02/09.aspx


The articles I shared below explain NASA and JPL’s involvement
below in addition our use of Tika, Nutch, and Lucene/Solr.

Cheers,
Chris

++
Chris Mattmann, Ph.D.
Chief Architect
Instrument Software and Science Data Systems Section (398)
NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
Office: 168-519, Mailstop: 168-527
Email: chris.a.mattm...@nasa.gov
WWW:  http://sunset.usc.edu/~mattmann/
++
Adjunct Associate Professor, Computer Science Department
University of Southern California, Los Angeles, CA 90089 USA
++




-Original Message-
From: Karl Wright daddy...@gmail.com
Reply-To: d...@nutch.apache.org d...@nutch.apache.org
Date: Wednesday, May 27, 2015 at 8:33 AM
To: dev@lucene.apache.org dev@lucene.apache.org
Cc: d...@tika.apache.org d...@tika.apache.org, d...@nutch.apache.org
d...@nutch.apache.org
Subject: Re: NASA/JPL's involvement in Memex: now public

The name Memex is, I think, already in use by a company that was bought
out by Autonomy some years ago, and thence by HP.  I presume this is
*not* what you are talking about?


Thanks,
Karl




On Wed, May 27, 2015 at 8:26 AM, Mattmann, Chris A (3980)
chris.a.mattm...@jpl.nasa.gov wrote:

Hey Folks,

Just wanted to share publicly some articles recently on NASA and
JPL’s involvement in Memex. It’s basically focused around Tika,
Nutch and Solr, so keep up the great work on all projects. A sampling
of the recent press/articles:

E. Landau. Deep Web Search May Help Scientists. NASA Jet Propulsion
Laboratory. Web
Feature. May 22, 2015.
http://www.jpl.nasa.gov/news/news.php?feature=4595
http://www.nasa.gov/jpl/deep-web-search-may-help-scientists



R. Chirgwin. JPL joins DARPA’s Memex project: Better search and indexing
for space stuff.
The Register -
UK. May 27, 2015.
http://www.theregister.co.uk/2015/05/27/jpl_joins_darpas_memex_project/



J. Lendino. NASA, DARPA collaborating on Deep Web search to analyze
spacecraft data.
ExtremeTech. May 26, 2015.
http://goo.gl/iRAFXC



M. Prigg. Nasa joins US government project to create ’Google for the dark
net’ that could
uncover cyber criminals, paedophiles and drug dealers in the online
underworld. Daily Mail -
UK. May 25, 2015.
http://goo.gl/xMj2hH



J. Wolman. Knee-deep in data. The Positive. May 27, 2015.
http://thepositive.com/knee-deep-in-data/



Scientists will benefit more from Deep Web Searches. GreenAtom. May 26,
2015.
http://goo.gl/UqDJAN



NASA to put science face on scary DARPA project. TRTWorld.com. May 26,
2015.
http://goo.gl/HJPv5O



Kitware Participates in DARPA Memex Program, Developing Software to
Address Complex
Search Problems. PR News, May 23, 2015.
http://www.prweb.com/releases/2015/05/prweb12741017.htm

Any questions let me know.

Cheers,
Chris









++
Chris Mattmann, Ph.D.
Chief Architect
Instrument Software and Science Data Systems Section (398)
NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
Office: 168-519, Mailstop: 168-527
Email: chris.a.mattm...@nasa.gov
WWW:  http://sunset.usc.edu/~mattmann/
++
Adjunct Associate Professor, Computer Science Department
University of Southern California, Los Angeles, CA 90089 USA
++









-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-7146) MiniSolrCloudCluster based tests can fail with ZooKeeperException NoNode for /live_nodes

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7146?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14560906#comment-14560906
 ] 

ASF subversion and git services commented on SOLR-7146:
---

Commit 1682003 from sha...@apache.org in branch 'dev/branches/branch_5x'
[ https://svn.apache.org/r1682003 ]

SOLR-7146: MiniSolrCloudCluster based tests can fail with ZooKeeperException 
NoNode for /live_nodes

 MiniSolrCloudCluster based tests can fail with ZooKeeperException NoNode for 
 /live_nodes
 

 Key: SOLR-7146
 URL: https://issues.apache.org/jira/browse/SOLR-7146
 Project: Solr
  Issue Type: Bug
  Components: SolrCloud, Tests
Reporter: Shalin Shekhar Mangar
Assignee: Shalin Shekhar Mangar
Priority: Minor
 Fix For: Trunk, 5.2

 Attachments: SOLR-7146.patch, SOLR-7146v2.patch


 MiniSolrCloudCluster based tests can fail with the following exception:
 {code}
 org.apache.solr.common.cloud.ZooKeeperException: 
   at 
 __randomizedtesting.SeedInfo.seed([3F3D838A8ADC9385:F153ADFBF163EC6D]:0)
   at 
 org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:463)
   at 
 org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:763)
   at 
 org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:752)
   at 
 org.apache.solr.cloud.MiniSolrCloudCluster.createCollection(MiniSolrCloudCluster.java:193)
   at 
 org.apache.solr.handler.component.TestTwoPhaseDistributedQuery.testNoExtraFieldsRequestedFromShardsInPhaseOne(TestTwoPhaseDistributedQuery.java:79)
   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877)
   at 
 com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
   at 
 org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
   at 
 org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
   at 
 com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
   at 
 org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
   at 
 org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
   at 
 org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
   at 
 com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   at 
 com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
   at 
 com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
   at 
 com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783)
   at 
 com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   at 
 com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
   at 
 org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
   at 
 org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
   at 
 com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
   at 
 com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
   at 
 

[jira] [Resolved] (SOLR-7146) MiniSolrCloudCluster based tests can fail with ZooKeeperException NoNode for /live_nodes

2015-05-27 Thread Shalin Shekhar Mangar (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-7146?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shalin Shekhar Mangar resolved SOLR-7146.
-

I increased the timeout to a minute and changed the logic to break immediately 
if numservers is reached.

Thanks Vamsee!

 MiniSolrCloudCluster based tests can fail with ZooKeeperException NoNode for 
 /live_nodes
 

 Key: SOLR-7146
 URL: https://issues.apache.org/jira/browse/SOLR-7146
 Project: Solr
  Issue Type: Bug
  Components: SolrCloud, Tests
Reporter: Shalin Shekhar Mangar
Assignee: Shalin Shekhar Mangar
Priority: Minor
 Fix For: Trunk, 5.3

 Attachments: SOLR-7146.patch, SOLR-7146v2.patch


 MiniSolrCloudCluster based tests can fail with the following exception:
 {code}
 org.apache.solr.common.cloud.ZooKeeperException: 
   at 
 __randomizedtesting.SeedInfo.seed([3F3D838A8ADC9385:F153ADFBF163EC6D]:0)
   at 
 org.apache.solr.client.solrj.impl.CloudSolrClient.connect(CloudSolrClient.java:463)
   at 
 org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:763)
   at 
 org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:752)
   at 
 org.apache.solr.cloud.MiniSolrCloudCluster.createCollection(MiniSolrCloudCluster.java:193)
   at 
 org.apache.solr.handler.component.TestTwoPhaseDistributedQuery.testNoExtraFieldsRequestedFromShardsInPhaseOne(TestTwoPhaseDistributedQuery.java:79)
   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1618)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:827)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:863)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:877)
   at 
 com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
   at 
 org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
   at 
 org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
   at 
 com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
   at 
 org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
   at 
 org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
   at 
 org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
   at 
 com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   at 
 com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
   at 
 com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
   at 
 com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:836)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:738)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:772)
   at 
 com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:783)
   at 
 com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   at 
 com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
   at 
 org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
   at 
 org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
   at 
 com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
   at 
 com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
   at 
 com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
   at 
 com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
   at 
 

[jira] [Updated] (LUCENE-6504) implement norms with random access API

2015-05-27 Thread Robert Muir (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-6504?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Muir updated LUCENE-6504:

Attachment: LUCENE-6504.patch

Here is a patch. We just use these methods to get signed values as a very 
simplistic compression. 

Generally speaking the performance seems ok, I think its the right tradeoff.

{noformat}
Report after iter 19:
Chart saved to out.png... (wd: /home/rmuir/workspace/util/src/python)
Task   QPS trunk  StdDev   QPS patch  StdDev
Pct diff
 LowTerm  983.09  (3.8%)  903.02  (4.6%)   
-8.1% ( -15% -0%)
HighSpanNear  158.40  (2.7%)  148.01  (2.0%)   
-6.6% ( -11% -   -1%)
OrNotHighMed  213.20  (3.9%)  199.50  (2.8%)   
-6.4% ( -12% -0%)
OrNotHighLow 1170.07  (2.7%) 1134.55  (2.4%)   
-3.0% (  -7% -2%)
 AndHighHigh   87.91  (1.9%)   86.21  (1.8%)   
-1.9% (  -5% -1%)
  IntNRQ8.64  (5.6%)8.48  (8.0%)   
-1.8% ( -14% -   12%)
  AndHighMed  123.04  (1.8%)  120.85  (1.7%)   
-1.8% (  -5% -1%)
  Fuzzy2   60.37  (1.3%)   59.35  (1.8%)   
-1.7% (  -4% -1%)
Wildcard   44.77  (3.2%)   44.06  (4.6%)   
-1.6% (  -9% -6%)
 MedSpanNear  150.07  (3.1%)  148.15  (2.6%)   
-1.3% (  -6% -4%)
 LowSpanNear   30.53  (1.2%)   30.15  (1.5%)   
-1.2% (  -3% -1%)
   LowPhrase   33.89  (2.1%)   33.49  (3.7%)   
-1.2% (  -6% -4%)
 Prefix3  210.10  (3.8%)  207.61  (5.3%)   
-1.2% (  -9% -8%)
  AndHighLow 1180.40  (2.1%) 1166.82  (2.5%)   
-1.2% (  -5% -3%)
 Respell   81.67  (1.5%)   81.41  (2.4%)   
-0.3% (  -4% -3%)
  Fuzzy1   97.84  (1.3%)   97.64  (1.7%)   
-0.2% (  -3% -2%)
 LowSloppyPhrase  120.00  (3.0%)  120.42  (3.0%)
0.4% (  -5% -6%)
   MedPhrase  263.96  (4.8%)  265.05  (7.1%)
0.4% ( -10% -   12%)
 MedSloppyPhrase   15.97  (4.5%)   16.08  (4.6%)
0.7% (  -8% -   10%)
 MedTerm  175.47  (4.4%)  177.64  (5.7%)
1.2% (  -8% -   11%)
  HighPhrase   17.80  (6.3%)   18.20  (9.3%)
2.2% ( -12% -   19%)
   OrHighMed   54.02  (6.8%)   56.18  (7.7%)
4.0% (  -9% -   19%)
   OrHighLow   52.35  (6.9%)   54.61  (7.7%)
4.3% (  -9% -   20%)
HighSloppyPhrase   12.85 (10.3%)   13.41 (11.9%)
4.4% ( -16% -   29%)
  OrHighHigh   25.34  (7.2%)   26.77  (8.3%)
5.6% (  -9% -   22%)
HighTerm  119.17  (4.7%)  128.45  (7.1%)
7.8% (  -3% -   20%)
OrHighNotLow  110.06  (6.4%)  119.06  (7.0%)
8.2% (  -4% -   23%)
OrHighNotMed   91.03  (6.1%)   98.91  (6.4%)
8.7% (  -3% -   22%)
   OrNotHighHigh   51.53  (5.6%)   56.04  (6.4%)
8.8% (  -3% -   21%)
   OrHighNotHigh   30.45  (5.7%)   33.27  (6.2%)
9.2% (  -2% -   22%)
{noformat}

 implement norms with random access API
 --

 Key: LUCENE-6504
 URL: https://issues.apache.org/jira/browse/LUCENE-6504
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Robert Muir
 Attachments: LUCENE-6504.patch


 We added this api in LUCENE-5729 but we never explored implementing norms 
 with it. These are generally the largest consumer of heap memory and often a 
 real hassle for users.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-7555) Display total space and available space in Admin

2015-05-27 Thread Eric Pugh (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-7555?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eric Pugh updated SOLR-7555:

Attachment: SOLR-7555-display_disk_space_v2.patch

Here is an updated patch.  I moved the method spaceAvailable and spaceTotal  to 
the abstract Directory class, and then implemented it for the various classes 
that extended it.   I don't love that this change touches the Lucene package, 
as I've never really gone deep there before, and it feels rather invasive!   

I also changed the format of the response to mimic the how the memory hands 
back both human formatted and raw data.

 Display total space and available space in Admin
 

 Key: SOLR-7555
 URL: https://issues.apache.org/jira/browse/SOLR-7555
 Project: Solr
  Issue Type: Improvement
  Components: web gui
Affects Versions: 5.1
Reporter: Eric Pugh
Assignee: Erik Hatcher
Priority: Minor
 Fix For: 5.2

 Attachments: SOLR-7555-display_disk_space.patch, 
 SOLR-7555-display_disk_space_v2.patch, SOLR-7555.patch


 Frequently I have access to the Solr Admin console, but not the underlying 
 server, and I'm curious how much space remains available.   This little patch 
 exposes total Volume size as well as the usable space remaining:
 !https://monosnap.com/file/VqlReekCFwpK6utI3lP18fbPqrGI4b.png!
 I'm not sure if this is the best place to put this, as every shard will share 
 the same data, so maybe it should be on the top level Dashboard?  Also not 
 sure what to call the fields! 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-7555) Display total space and available space in Admin

2015-05-27 Thread Eric Pugh (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7555?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561441#comment-14561441
 ] 

Eric Pugh commented on SOLR-7555:
-

[~mariusneo]this was perfect!

 Display total space and available space in Admin
 

 Key: SOLR-7555
 URL: https://issues.apache.org/jira/browse/SOLR-7555
 Project: Solr
  Issue Type: Improvement
  Components: web gui
Affects Versions: 5.1
Reporter: Eric Pugh
Assignee: Erik Hatcher
Priority: Minor
 Fix For: 5.2

 Attachments: SOLR-7555-display_disk_space.patch, 
 SOLR-7555-display_disk_space_v2.patch, SOLR-7555.patch


 Frequently I have access to the Solr Admin console, but not the underlying 
 server, and I'm curious how much space remains available.   This little patch 
 exposes total Volume size as well as the usable space remaining:
 !https://monosnap.com/file/VqlReekCFwpK6utI3lP18fbPqrGI4b.png!
 I'm not sure if this is the best place to put this, as every shard will share 
 the same data, so maybe it should be on the top level Dashboard?  Also not 
 sure what to call the fields! 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Reopened] (SOLR-6820) The sync on the VersionInfo bucket in DistributedUpdateProcesser#addDocument appears to be a large bottleneck when using replication.

2015-05-27 Thread Timothy Potter (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-6820?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Timothy Potter reopened SOLR-6820:
--

Copy-and-paste error in the configset solrconfig.xml files, the int parameter 
is missing the name!

{code}
updateLog
  str name=dir${solr.ulog.dir:}/str
  int name=${solr.ulog.numVersionBuckets:65536}/int
/updateLog
{code}

but should be:

{code}
updateLog
  str name=dir${solr.ulog.dir:}/str
  int name=numVersionBuckets${solr.ulog.numVersionBuckets:65536}/int
/updateLog
{code}

It doesn't look like this causes any failures but looks bad.

 The sync on the VersionInfo bucket in DistributedUpdateProcesser#addDocument 
 appears to be a large bottleneck when using replication.
 -

 Key: SOLR-6820
 URL: https://issues.apache.org/jira/browse/SOLR-6820
 Project: Solr
  Issue Type: Sub-task
  Components: SolrCloud
Reporter: Mark Miller
Assignee: Timothy Potter
 Fix For: Trunk, 5.2

 Attachments: SOLR-6820.patch, threads.png






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: Building the first RC for Lucene/Solr 5.2.0

2015-05-27 Thread Timothy Potter
Hi Anshum,

I re-opened SOLR-6820 due to a minor issue with a parameter in
solrconfig.xml missing the name attribute. It doesn't cause any
failures, but looks bad ... so I don't think it warrants a re-spin but
if we go to another RC, then I'll get that in ... if not, I'll close
SOLR-6820 and open a separate ticket for that problem.

Cheers,
Tim

On Wed, May 27, 2015 at 9:52 AM, Anshum Gupta ans...@anshumgupta.net wrote:
 FYI, I've started the process to build the first RC for Lucene/Solr 5.2.0

 --
 Anshum Gupta

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: Hmm, bug (schemaless/schema api)?

2015-05-27 Thread Steve Rowe

 On May 27, 2015, at 1:40 PM, Gus Heck gus.h...@gmail.com wrote:
 
 What you say is what I had generally understood to be the case in the past 
 with regular static schemas, but I had hoped something newer and more 
 flexible was implied with the advent of an API, and the schemaless 
 configuration. Although there is a caveat at the top of the wiki page, 
 statements like 
 
 When modifying the schema with the REST API, a core reload will 
 automatically occur in order for the changes to be available immediately. 
 
 had given me hope. Guess you haven't got that far after all.

Thanks, I added a(nother) warning to that sentence.

Steve
-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: Building the first RC for Lucene/Solr 5.2.0

2015-05-27 Thread Anshum Gupta
Sure, sounds good to me. I'm running the smoketester now so the RC should
be out for vote right after.

On Wed, May 27, 2015 at 11:41 AM, Timothy Potter thelabd...@gmail.com
wrote:

 Hi Anshum,

 I re-opened SOLR-6820 due to a minor issue with a parameter in
 solrconfig.xml missing the name attribute. It doesn't cause any
 failures, but looks bad ... so I don't think it warrants a re-spin but
 if we go to another RC, then I'll get that in ... if not, I'll close
 SOLR-6820 and open a separate ticket for that problem.

 Cheers,
 Tim

 On Wed, May 27, 2015 at 9:52 AM, Anshum Gupta ans...@anshumgupta.net
 wrote:
  FYI, I've started the process to build the first RC for Lucene/Solr 5.2.0
 
  --
  Anshum Gupta

 -
 To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
 For additional commands, e-mail: dev-h...@lucene.apache.org




-- 
Anshum Gupta


[jira] [Commented] (SOLR-7344) Use two thread pools, one for internal requests and one for external, to avoid distributed deadlock and decrease the number of threads that need to be created.

2015-05-27 Thread Hrishikesh Gadre (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7344?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561532#comment-14561532
 ] 

Hrishikesh Gadre commented on SOLR-7344:


[~yo...@apache.org]

I have an alternate proposal similar to your earlier idea of having a logical 
partition of a jetty worker pool without requiring a separate endpoint. I 
believe this would also solve the problem of request forwarding.

The idea here is to implement a custom Servlet filter similar to  [Jetty 
QoSFilter|http://www.eclipse.org/jetty/documentation/current/qos-filter.html] 
which would reserve a certain capacity of the worker thread-pool for internal 
and external requests. The advantages of this approach are as follows,

- We will still get to use the thread-per-request model when the number of 
requests (of a specific type) are less than the capacity reserved for that 
request type. All the additional requests will be suspended until atleast one 
thread is available for processing.
- We will not require extensive changes to the SolrDispatcherFilter code 
thereby reducing the possibility of introducing new bugs.
- Since this new filter needs to be added to the solr webapp, folks can opt out 
if they want (Please see the earlier comment from [~andyetitmoves]).

Note even for this approach, we need to implement a way to identify an internal 
request (e.g. adding an extra request header/param). Now regarding the request 
forwarding, I have following change in mind. Let's assume that a client C has 
sent a request to node A for a collection/core hosted by node B. 

- The initial request from client C will be categorized as an external 
request.
- Node A will figure out that it does not host the requested collection/core. 
It will figure out the correct Solr server (node B in this case) and send 
another request to node B. Here we will ensure that this request is also 
categorized as an external request (i.e. not to add the request header/param 
even though its a server-server communication).
- Node B will process the forwarded request as if it were an external 
request and return results to node A.
- Node A will return results to client C.



 Use two thread pools, one for internal requests and one for external, to 
 avoid distributed deadlock and decrease the number of threads that need to be 
 created.
 ---

 Key: SOLR-7344
 URL: https://issues.apache.org/jira/browse/SOLR-7344
 Project: Solr
  Issue Type: Improvement
  Components: SolrCloud
Reporter: Mark Miller
 Attachments: SOLR-7344.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-7555) Display total space and available space in Admin

2015-05-27 Thread Upayavira (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7555?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561545#comment-14561545
 ] 

Upayavira commented on SOLR-7555:
-

Extending the Directory class seems like a pretty major step - one that should 
get far wider coverage in its own ticket, I'd say.

However, a simpler way to do it would be to have a DiskSpaceAwareDirectory 
interface that provides the necessary methods. Then, the API can only offer 
disk usage values from Directory objects that implement that interface.

Dunno if that pattern is used at all in Lucene/Solr though.

 Display total space and available space in Admin
 

 Key: SOLR-7555
 URL: https://issues.apache.org/jira/browse/SOLR-7555
 Project: Solr
  Issue Type: Improvement
  Components: web gui
Affects Versions: 5.1
Reporter: Eric Pugh
Assignee: Erik Hatcher
Priority: Minor
 Fix For: 5.2

 Attachments: SOLR-7555-display_disk_space.patch, 
 SOLR-7555-display_disk_space_v2.patch, SOLR-7555.patch


 Frequently I have access to the Solr Admin console, but not the underlying 
 server, and I'm curious how much space remains available.   This little patch 
 exposes total Volume size as well as the usable space remaining:
 !https://monosnap.com/file/VqlReekCFwpK6utI3lP18fbPqrGI4b.png!
 I'm not sure if this is the best place to put this, as every shard will share 
 the same data, so maybe it should be on the top level Dashboard?  Also not 
 sure what to call the fields! 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Building the first RC for Lucene/Solr 5.2.0

2015-05-27 Thread Anshum Gupta
FYI, I've started the process to build the first RC for Lucene/Solr 5.2.0

-- 
Anshum Gupta


[jira] [Commented] (LUCENE-6487) Add WGS84 capability to geo3d support

2015-05-27 Thread Karl Wright (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6487?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561176#comment-14561176
 ] 

Karl Wright commented on LUCENE-6487:
-

Simple round-trip tests are trivial, but I won't get to one today, and I agree 
that it is not a blocker for commit.  We can create a separate ticket if you go 
ahead and merge.

re doubles vs. floats: Well, the actual (x,y,z) value needs to be accurate to 
within 1e-12 in order to properly specify various planes etc throughout the 
module.  So I'd be cautious about deliberately reducing precision, since 
geopoint extends vector and plane does too.  We can experiment but...  once 
again, I think a separate ticket would be in order.

 Add WGS84 capability to geo3d support
 -

 Key: LUCENE-6487
 URL: https://issues.apache.org/jira/browse/LUCENE-6487
 Project: Lucene - Core
  Issue Type: Improvement
  Components: modules/spatial
Reporter: Karl Wright
 Attachments: LUCENE-6487.patch, LUCENE-6487.patch, LUCENE-6487.patch, 
 LUCENE-6487.patch


 WGS84 compatibility has been requested for geo3d.  This involves working with 
 an ellipsoid rather than a unit sphere.  The general formula for an ellipsoid 
 is:
 x^2/a^2 + y^2/b^2 + z^2/c^2 = 1



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-7590) Finish and improve MDC logging support.

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7590?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561136#comment-14561136
 ] 

ASF subversion and git services commented on SOLR-7590:
---

Commit 1682057 from [~markrmil...@gmail.com] in branch 'dev/branches/branch_5x'
[ https://svn.apache.org/r1682057 ]

SOLR-7590: Java8 code - Java7 for branch_5x.

 Finish and improve MDC logging support.
 ---

 Key: SOLR-7590
 URL: https://issues.apache.org/jira/browse/SOLR-7590
 Project: Solr
  Issue Type: Improvement
Reporter: Mark Miller
Assignee: Mark Miller
 Fix For: Trunk, 5.3

 Attachments: SOLR-7950.patch, SOLR-7950.patch, SOLR-7950.patch, 
 SOLR-7950.patch, SOLR-7950.patch, SOLR-7950.patch, SOLR-7950.patch, 
 SOLR-7950.patch, SOLR-7950.patch, SOLR-7950.patch, SOLR-7950.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-7361) Main Jetty thread blocked by core loading delays HTTP listener from binding if core loading is slow

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7361?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561158#comment-14561158
 ] 

ASF subversion and git services commented on SOLR-7361:
---

Commit 1682060 from [~markrmil...@gmail.com] in branch 'dev/trunk'
[ https://svn.apache.org/r1682060 ]

SOLR-7361: Slow loading SolrCores should not hold up all other SolrCores that 
have finished loading from serving requests.

 Main Jetty thread blocked by core loading delays HTTP listener from binding 
 if core loading is slow
 ---

 Key: SOLR-7361
 URL: https://issues.apache.org/jira/browse/SOLR-7361
 Project: Solr
  Issue Type: Bug
  Components: SolrCloud
Reporter: Timothy Potter
Assignee: Mark Miller
 Fix For: Trunk, 5.2

 Attachments: SOLR-7361.patch, SOLR-7361.patch, SOLR-7361.patch, 
 SOLR-7361.patch, SOLR-7361.patch, SOLR-7361.patch, SOLR-7361.patch, 
 SOLR-7361.patch


 During server startup, the CoreContainer uses an ExecutorService to load 
 cores in multiple back-ground threads but then blocks until cores are loaded, 
 see: CoreContainer#load around line 290 on trunk (invokeAll). From the 
 JavaDoc on that method, we have:
 {quote}
 Executes the given tasks, returning a list of Futures holding their status 
 and results when all complete. Future.isDone() is true for each element of 
 the returned list.
 {quote}
 In other words, this is a blocking call.
 This delays the Jetty HTTP listener from binding and accepting requests until 
 all cores are loaded. Do we need to block the main thread?
 Also, prior to this happening, the node is registered as a live node in ZK, 
 which makes it a candidate for receiving requests from the Overseer, such as 
 to service a create collection request. The problem of course is that the 
 node listed in /live_nodes isn't accepting requests yet. So we either need to 
 unblock the main thread during server loading or maybe wait longer before we 
 register as a live node ... not sure which is the better way forward?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-7361) Main Jetty thread blocked by core loading delays HTTP listener from binding if core loading is slow

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7361?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561175#comment-14561175
 ] 

ASF subversion and git services commented on SOLR-7361:
---

Commit 1682065 from [~markrmil...@gmail.com] in branch 'dev/branches/branch_5x'
[ https://svn.apache.org/r1682065 ]

SOLR-7361: Slow loading SolrCores should not hold up all other SolrCores that 
have finished loading from serving requests.

 Main Jetty thread blocked by core loading delays HTTP listener from binding 
 if core loading is slow
 ---

 Key: SOLR-7361
 URL: https://issues.apache.org/jira/browse/SOLR-7361
 Project: Solr
  Issue Type: Bug
  Components: SolrCloud
Reporter: Timothy Potter
Assignee: Mark Miller
 Fix For: Trunk, 5.2

 Attachments: SOLR-7361.patch, SOLR-7361.patch, SOLR-7361.patch, 
 SOLR-7361.patch, SOLR-7361.patch, SOLR-7361.patch, SOLR-7361.patch, 
 SOLR-7361.patch


 During server startup, the CoreContainer uses an ExecutorService to load 
 cores in multiple back-ground threads but then blocks until cores are loaded, 
 see: CoreContainer#load around line 290 on trunk (invokeAll). From the 
 JavaDoc on that method, we have:
 {quote}
 Executes the given tasks, returning a list of Futures holding their status 
 and results when all complete. Future.isDone() is true for each element of 
 the returned list.
 {quote}
 In other words, this is a blocking call.
 This delays the Jetty HTTP listener from binding and accepting requests until 
 all cores are loaded. Do we need to block the main thread?
 Also, prior to this happening, the node is registered as a live node in ZK, 
 which makes it a candidate for receiving requests from the Overseer, such as 
 to service a create collection request. The problem of course is that the 
 node listed in /live_nodes isn't accepting requests yet. So we either need to 
 unblock the main thread during server loading or maybe wait longer before we 
 register as a live node ... not sure which is the better way forward?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Hmm, bug?

2015-05-27 Thread Gus Heck
So I did the following (seemingly) reasonable set of steps...


   1. Set up a Solr Cloud instance with mutable managed schema (data driven
   config)
   2. Tossed a couple docs in just to see it work.
   3. Tried to sort on 'title', realized title had been guessed as
   multivalued.
   4. Changed the field to type=string (as opposed to strings) using schema
   api
   5. Tried to sort on title

The result is

   -

   trace: java.lang.IllegalStateException: unexpected docvalues
type SORTED_SET for field 'title' (expected=SORTED). Use
UninvertingReader or index with docvalues.\n\tat
org.apache.lucene.index.DocValues.checkField(DocValues.java:208)

   - Resending the curl command to re-index the docs hangs

However, I have not played with schemaless or the schema api before, so
maybe I've missed something, and my steps are not reasonable, and thus this
sanity check before I file a bug.

-Gus

-- 
http://www.the111shift.com


[jira] [Commented] (LUCENE-6481) Improve GeoPointField type to only visit high precision boundary terms

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6481?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561056#comment-14561056
 ] 

ASF subversion and git services commented on LUCENE-6481:
-

Commit 1682043 from [~mikemccand] in branch 'dev/branches/LUCENE-6481'
[ https://svn.apache.org/r1682043 ]

LUCENE-6481: add nocommit about query slowness for large bboxes; improve 
javadocs

 Improve GeoPointField type to only visit high precision boundary terms 
 ---

 Key: LUCENE-6481
 URL: https://issues.apache.org/jira/browse/LUCENE-6481
 Project: Lucene - Core
  Issue Type: Improvement
  Components: core/index
Reporter: Nicholas Knize
 Attachments: LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481.patch, 
 LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481_WIP.patch


 Current GeoPointField [LUCENE-6450 | 
 https://issues.apache.org/jira/browse/LUCENE-6450] computes a set of ranges 
 along the space-filling curve that represent a provided bounding box.  This 
 determines which terms to visit in the terms dictionary and which to skip. 
 This is suboptimal for large bounding boxes as we may end up visiting all 
 terms (which could be quite large). 
 This incremental improvement is to improve GeoPointField to only visit high 
 precision terms in boundary ranges and use the postings list for ranges that 
 are completely within the target bounding box.
 A separate improvement is to switch over to auto-prefix and build an 
 Automaton representing the bounding box.  That can be tracked in a separate 
 issue.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4506) [solr4.0.0] many index.{date} dir in replication node

2015-05-27 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4506?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561089#comment-14561089
 ] 

Mark Miller commented on SOLR-4506:
---

Fire away - fell off my radar - can't remember 2013 :)

 [solr4.0.0] many index.{date} dir in replication node 
 --

 Key: SOLR-4506
 URL: https://issues.apache.org/jira/browse/SOLR-4506
 Project: Solr
  Issue Type: Bug
  Components: SolrCloud
Affects Versions: 4.0
 Environment: the solr4.0 runs on suse11.
 mem:32G
 cpu:16 cores
Reporter: zhuojunjian
Assignee: Timothy Potter
Priority: Minor
 Fix For: 4.9, Trunk

   Original Estimate: 12h
  Remaining Estimate: 12h

 in our test,we used solrcloud feature in solr4.0(version detail 
 :4.0.0.2012.10.06.03.04.33).
 the solrcloud configuration is 3 shards and 2 replications each shard.
 we found that there are over than 25 dirs which named index.{date} in one 
 replication node belonging to shard 3. 
 for example:
 index.2013021725864  index.20130218012211880  index.20130218015714713  
 index.20130218023101958  index.20130218030424083  tlog
 index.20130218005648324  index.20130218012751078  index.20130218020141293  
 the issue seems like SOLR-1781. but it is fixed in 4.0-BETA,5.0. 
 so is solr4.0 ? if it is fixed too in solr4.0, why we find the issue again ?
 what can I do?   



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-4506) [solr4.0.0] many index.{date} dir in replication node

2015-05-27 Thread Mark Miller (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4506?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Miller updated SOLR-4506:
--
Assignee: Timothy Potter  (was: Mark Miller)

 [solr4.0.0] many index.{date} dir in replication node 
 --

 Key: SOLR-4506
 URL: https://issues.apache.org/jira/browse/SOLR-4506
 Project: Solr
  Issue Type: Bug
  Components: SolrCloud
Affects Versions: 4.0
 Environment: the solr4.0 runs on suse11.
 mem:32G
 cpu:16 cores
Reporter: zhuojunjian
Assignee: Timothy Potter
Priority: Minor
 Fix For: 4.9, Trunk

   Original Estimate: 12h
  Remaining Estimate: 12h

 in our test,we used solrcloud feature in solr4.0(version detail 
 :4.0.0.2012.10.06.03.04.33).
 the solrcloud configuration is 3 shards and 2 replications each shard.
 we found that there are over than 25 dirs which named index.{date} in one 
 replication node belonging to shard 3. 
 for example:
 index.2013021725864  index.20130218012211880  index.20130218015714713  
 index.20130218023101958  index.20130218030424083  tlog
 index.20130218005648324  index.20130218012751078  index.20130218020141293  
 the issue seems like SOLR-1781. but it is fixed in 4.0-BETA,5.0. 
 so is solr4.0 ? if it is fixed too in solr4.0, why we find the issue again ?
 what can I do?   



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6504) implement norms with random access API

2015-05-27 Thread Uwe Schindler (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6504?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561095#comment-14561095
 ] 

Uwe Schindler commented on LUCENE-6504:
---

bq. Its not unrelated, i did extensive benchmarking. That optimization is a 
trap, allowing too many low level implementations (3) once the index gets large.

OK. That's fine. I never liked that offset=0 specialization, i think back at 
that time we felt it might be a good idea. I don't know the issue number 
anymore.

You also removed the negative check in the new Multi implementation, but I 
think this was done because of performance, too. I am not happy with that, 
because it no longer throws Exception if you try to access stuff off-slice. 
Maybe we can add an assert instead?

 implement norms with random access API
 --

 Key: LUCENE-6504
 URL: https://issues.apache.org/jira/browse/LUCENE-6504
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Robert Muir
 Attachments: LUCENE-6504.patch


 We added this api in LUCENE-5729 but we never explored implementing norms 
 with it. These are generally the largest consumer of heap memory and often a 
 real hassle for users.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-NightlyTests-trunk - Build # 694 - Still Failing

2015-05-27 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-trunk/694/

4 tests failed.
REGRESSION:  org.apache.solr.cloud.hdfs.HdfsChaosMonkeySafeLeaderTest.test

Error Message:
The Monkey ran for over 20 seconds and no jetties were stopped - this is worth 
investigating!

Stack Trace:
java.lang.AssertionError: The Monkey ran for over 20 seconds and no jetties 
were stopped - this is worth investigating!
at 
__randomizedtesting.SeedInfo.seed([BF0290866116A66E:3756AF5CCFEACB96]:0)
at org.junit.Assert.fail(Assert.java:93)
at org.apache.solr.cloud.ChaosMonkey.stopTheMonkey(ChaosMonkey.java:537)
at 
org.apache.solr.cloud.ChaosMonkeySafeLeaderTest.test(ChaosMonkeySafeLeaderTest.java:152)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:483)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:836)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:872)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:886)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:960)
at 
org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:935)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:845)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:747)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:781)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 

Re: svn commit: r1682031 [1/2] - in /lucene/dev/branches/branch_5x: ./ solr/ solr/core/ solr/core/src/java/org/apache/solr/ solr/core/src/java/org/apache/solr/cloud/ solr/core/src/java/org/apache/solr

2015-05-27 Thread Steve Rowe
Java7 compilation borked?:

   [javac] Compiling 807 source files to 
/var/lib/jenkins/jobs/Solr-core-tests-5x-Java7/workspace/solr/build/solr-core/classes/java
   [javac] 
/var/lib/jenkins/jobs/Solr-core-tests-5x-Java7/workspace/solr/core/src/java/org/apache/solr/logging/MDCLoggingContext.java:26:
 error: package java.util.function does not exist
   [javac] import java.util.function.Supplier;
   [javac]  ^
   [javac] 
/var/lib/jenkins/jobs/Solr-core-tests-5x-Java7/workspace/solr/core/src/java/org/apache/solr/logging/MDCLoggingContext.java:42:
 error: cannot find symbol
   [javac]   private static ThreadLocalInteger CALL_DEPTH = 
ThreadLocal.withInitial(new SupplierInteger() {
   [javac]  
  ^
   [javac]   symbol:   class Supplier
   [javac]   location: class MDCLoggingContext
   [javac] Note: Some input files use or override a deprecated API.
   [javac] Note: Recompile with -Xlint:deprecation for details.
   [javac] Note: Some input files use unchecked or unsafe operations.
   [javac] Note: Recompile with -Xlint:unchecked for details.
   [javac] 2 errors

Steve

 On May 27, 2015, at 9:58 AM, markrmil...@apache.org wrote:
 
 Author: markrmiller
 Date: Wed May 27 13:58:30 2015
 New Revision: 1682031
 
 URL: http://svn.apache.org/r1682031
 Log:
 SOLR-7590: Finish and improve MDC context logging support.
 
 Added:

 lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/logging/MDCLoggingContext.java
  - copied unchanged from r1682020, 
 lucene/dev/trunk/solr/core/src/java/org/apache/solr/logging/MDCLoggingContext.java
 Removed:

 lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/SolrLogFormatter.java

 lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/logging/MDCUtils.java
 Modified:
lucene/dev/branches/branch_5x/   (props changed)
lucene/dev/branches/branch_5x/solr/   (props changed)
lucene/dev/branches/branch_5x/solr/CHANGES.txt   (contents, props changed)
lucene/dev/branches/branch_5x/solr/core/   (props changed)

 lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/cloud/ElectionContext.java

 lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/cloud/OverseerCollectionProcessor.java

 lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/cloud/RecoveryStrategy.java

 lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/cloud/SyncStrategy.java

 lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/cloud/ZkController.java

 lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/core/CoreContainer.java

 lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/core/CoreDescriptor.java

 lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/core/SolrCore.java

 lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/core/SolrCores.java

 lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/core/ZkContainer.java

 lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/servlet/HttpSolrCall.java

 lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/update/DefaultSolrCoreState.java

 lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/update/PeerSync.java
lucene/dev/branches/branch_5x/solr/core/src/test-files/log4j.properties
lucene/dev/branches/branch_5x/solr/server/   (props changed)
lucene/dev/branches/branch_5x/solr/server/resources/log4j.properties
lucene/dev/branches/branch_5x/solr/solrj/   (props changed)

 lucene/dev/branches/branch_5x/solr/solrj/src/java/org/apache/solr/common/util/ExecutorUtil.java
lucene/dev/branches/branch_5x/solr/solrj/src/test-files/log4j.properties
lucene/dev/branches/branch_5x/solr/test-framework/   (props changed)

 lucene/dev/branches/branch_5x/solr/test-framework/src/java/org/apache/solr/SolrTestCaseJ4.java

 lucene/dev/branches/branch_5x/solr/test-framework/src/java/org/apache/solr/cloud/AbstractFullDistribZkTestBase.java
 
 Modified: lucene/dev/branches/branch_5x/solr/CHANGES.txt
 URL: 
 http://svn.apache.org/viewvc/lucene/dev/branches/branch_5x/solr/CHANGES.txt?rev=1682031r1=1682030r2=1682031view=diff
 ==
 --- lucene/dev/branches/branch_5x/solr/CHANGES.txt (original)
 +++ lucene/dev/branches/branch_5x/solr/CHANGES.txt Wed May 27 13:58:30 2015
 @@ -45,6 +45,8 @@ Other Changes
 * SOLR-7146: MiniSolrCloudCluster based tests can fail with 
 ZooKeeperException NoNode for /live_nodes.
   (Vamsee Yarlagadda via shalin)
 
 +* SOLR-7590: Finish and improve MDC context logging support. (Mark Miller)  
 +
 ==  5.2.0 ==
 
 Consult the LUCENE_CHANGES.txt file for additional, low level, changes in 
 this release
 
 Modified: 
 

[jira] [Commented] (SOLR-7590) Finish and improve MDC logging support.

2015-05-27 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7590?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561131#comment-14561131
 ] 

Mark Miller commented on SOLR-7590:
---

Looks like 5x is failing because the Supplier code is Java 8. Still trying to 
figure out how this appears to work on my dev machine even when running with a 
java 7 jvm...

 Finish and improve MDC logging support.
 ---

 Key: SOLR-7590
 URL: https://issues.apache.org/jira/browse/SOLR-7590
 Project: Solr
  Issue Type: Improvement
Reporter: Mark Miller
Assignee: Mark Miller
 Fix For: Trunk, 5.3

 Attachments: SOLR-7950.patch, SOLR-7950.patch, SOLR-7950.patch, 
 SOLR-7950.patch, SOLR-7950.patch, SOLR-7950.patch, SOLR-7950.patch, 
 SOLR-7950.patch, SOLR-7950.patch, SOLR-7950.patch, SOLR-7950.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6481) Improve GeoPointField type to only visit high precision boundary terms

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6481?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561044#comment-14561044
 ] 

ASF subversion and git services commented on LUCENE-6481:
-

Commit 1682036 from [~mikemccand] in branch 'dev/branches/LUCENE-6481'
[ https://svn.apache.org/r1682036 ]

LUCENE-6481: restrict random test cases to 'smallish' bbox; switch static 
factory for polygon query to ctor

 Improve GeoPointField type to only visit high precision boundary terms 
 ---

 Key: LUCENE-6481
 URL: https://issues.apache.org/jira/browse/LUCENE-6481
 Project: Lucene - Core
  Issue Type: Improvement
  Components: core/index
Reporter: Nicholas Knize
 Attachments: LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481.patch, 
 LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481_WIP.patch


 Current GeoPointField [LUCENE-6450 | 
 https://issues.apache.org/jira/browse/LUCENE-6450] computes a set of ranges 
 along the space-filling curve that represent a provided bounding box.  This 
 determines which terms to visit in the terms dictionary and which to skip. 
 This is suboptimal for large bounding boxes as we may end up visiting all 
 terms (which could be quite large). 
 This incremental improvement is to improve GeoPointField to only visit high 
 precision terms in boundary ranges and use the postings list for ranges that 
 are completely within the target bounding box.
 A separate improvement is to switch over to auto-prefix and build an 
 Automaton representing the bounding box.  That can be tracked in a separate 
 issue.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Comment Edited] (LUCENE-6504) implement norms with random access API

2015-05-27 Thread Uwe Schindler (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6504?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561095#comment-14561095
 ] 

Uwe Schindler edited comment on LUCENE-6504 at 5/27/15 2:51 PM:


bq. Its not unrelated, i did extensive benchmarking. That optimization is a 
trap, allowing too many low level implementations (3) once the index gets large.

OK. That's fine. I never liked that offset=0 specialization, i think back at 
that time we felt it might be a good idea. I don't know the issue number 
anymore.

You also removed the negative check in the new Multi implementation; I think 
this was done because of performance, too, right? I am not happy with that, 
because it no longer throws Exception if you try to access stuff off-slice. 
Maybe we can add an assert instead?


was (Author: thetaphi):
bq. Its not unrelated, i did extensive benchmarking. That optimization is a 
trap, allowing too many low level implementations (3) once the index gets large.

OK. That's fine. I never liked that offset=0 specialization, i think back at 
that time we felt it might be a good idea. I don't know the issue number 
anymore.

You also removed the negative check in the new Multi implementation, but I 
think this was done because of performance, too. I am not happy with that, 
because it no longer throws Exception if you try to access stuff off-slice. 
Maybe we can add an assert instead?

 implement norms with random access API
 --

 Key: LUCENE-6504
 URL: https://issues.apache.org/jira/browse/LUCENE-6504
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Robert Muir
 Attachments: LUCENE-6504.patch


 We added this api in LUCENE-5729 but we never explored implementing norms 
 with it. These are generally the largest consumer of heap memory and often a 
 real hassle for users.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6504) implement norms with random access API

2015-05-27 Thread Uwe Schindler (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6504?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561100#comment-14561100
 ] 

Uwe Schindler commented on LUCENE-6504:
---

Sorry, the assert is there. Cancel my comment!

 implement norms with random access API
 --

 Key: LUCENE-6504
 URL: https://issues.apache.org/jira/browse/LUCENE-6504
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Robert Muir
 Attachments: LUCENE-6504.patch


 We added this api in LUCENE-5729 but we never explored implementing norms 
 with it. These are generally the largest consumer of heap memory and often a 
 real hassle for users.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6504) implement norms with random access API

2015-05-27 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6504?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561040#comment-14561040
 ] 

Robert Muir commented on LUCENE-6504:
-

{quote}
Before/now, AFAIK norms take 1 byte per field per doc of heap. I looked over 
the patch briefly; does this essentially put norms off-heap?
{quote}

In the worst case. Currently they are compressed with bitpacking and other 
tricks to try to be reasonable. But what was missing all along was a random 
access api in Directory so that this can just be MappedByteBuffer.get(long) 
(see linked issue and justification). If you want them to be in heap memory, 
use fileswitchdirectory and ramdirectory.

{quote}
Does this patch also bring in accurate norms or is something else required to 
enable actually utilize that?
{quote}

You have 1 byte norms because your chosen similarity squashes to that, but the 
interface between similarity and indexwriter is long since lucene 4 and all 
codecs test and support that.


 implement norms with random access API
 --

 Key: LUCENE-6504
 URL: https://issues.apache.org/jira/browse/LUCENE-6504
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Robert Muir
 Attachments: LUCENE-6504.patch


 We added this api in LUCENE-5729 but we never explored implementing norms 
 with it. These are generally the largest consumer of heap memory and often a 
 real hassle for users.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Solr-Artifacts-5.x - Build # 844 - Failure

2015-05-27 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Solr-Artifacts-5.x/844/

No tests ran.

Build Log:
[...truncated 12816 lines...]
[javac] Compiling 807 source files to 
/x1/jenkins/jenkins-slave/workspace/Solr-Artifacts-5.x/solr/build/solr-core/classes/java
[javac] 
/x1/jenkins/jenkins-slave/workspace/Solr-Artifacts-5.x/solr/core/src/java/org/apache/solr/logging/MDCLoggingContext.java:26:
 error: package java.util.function does not exist
[javac] import java.util.function.Supplier;
[javac]  ^
[javac] 
/x1/jenkins/jenkins-slave/workspace/Solr-Artifacts-5.x/solr/core/src/java/org/apache/solr/logging/MDCLoggingContext.java:42:
 error: cannot find symbol
[javac]   private static ThreadLocalInteger CALL_DEPTH = 
ThreadLocal.withInitial(new SupplierInteger() {
[javac] 
   ^
[javac]   symbol:   class Supplier
[javac]   location: class MDCLoggingContext
[javac] Note: Some input files use or override a deprecated API.
[javac] Note: Recompile with -Xlint:deprecation for details.
[javac] Note: Some input files use unchecked or unsafe operations.
[javac] Note: Recompile with -Xlint:unchecked for details.
[javac] 2 errors

BUILD FAILED
/x1/jenkins/jenkins-slave/workspace/Solr-Artifacts-5.x/solr/build.xml:510: The 
following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Solr-Artifacts-5.x/solr/common-build.xml:389:
 The following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Solr-Artifacts-5.x/lucene/common-build.xml:523:
 The following error occurred while executing this line:
/x1/jenkins/jenkins-slave/workspace/Solr-Artifacts-5.x/lucene/common-build.xml:1945:
 Compile failed; see the compiler error output for details.

Total time: 1 minute 26 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Sending artifact delta relative to Solr-Artifacts-5.x #843
Archived 3 artifacts
Archive block size is 32768
Received 0 blocks and 38101191 bytes
Compression is 0.0%
Took 1 min 12 sec
Publishing Javadoc
Email was triggered for: Failure
Sending email for trigger: Failure



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Commented] (SOLR-5386) Solr hangs on spellcheck.maxCollationTries

2015-05-27 Thread Bob Hastings (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-5386?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561035#comment-14561035
 ] 

Bob Hastings commented on SOLR-5386:


I am having the same problem with Solr 4.6.0.  The lock is occurring in 
SolrCore.getSearcher().  For me at startup both the 
[coreLoadExecutor-3-thread-1] and [searcherExecutor-4-thread-1] are blocked on 
the searhcerLock.wait() call line 1546 for SolrCore for release 4.6.0.  My 
guess is the synchronized block needs review.  I'll to more checking and let 
you know what I find.

BTW this happens every time this line is added to the /select RequestHandler:

str name=spellcheck.maxCollationTries5/str

Whenever the value is greater than zero.

 Solr hangs on spellcheck.maxCollationTries
 --

 Key: SOLR-5386
 URL: https://issues.apache.org/jira/browse/SOLR-5386
 Project: Solr
  Issue Type: Bug
  Components: spellchecker
Affects Versions: 4.4, 4.5
Reporter: Jeroen Steggink
  Labels: collate, maxCollationTries, spellcheck
 Attachments: threaddump.log


 When spellcheck.maxCollationTries is set (0) Solr hangs in combination with 
 that requestHandler set to default=true.
 When I make another requestHandler default, one without the 
 maxCollationTries, all requestHandlers work just fine.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6504) implement norms with random access API

2015-05-27 Thread Robert Muir (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6504?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561043#comment-14561043
 ] 

Robert Muir commented on LUCENE-6504:
-

{quote}
There are some unrelated changes in ByteBufferIndexInput: You removed the 
offset=0 implementation. Did you find out that this optimization for offset=0 
brings no additional performance?
{quote}

Its not unrelated, i did extensive benchmarking. That optimization is a trap, 
allowing too many low level implementations (3) once the index gets large. 

 implement norms with random access API
 --

 Key: LUCENE-6504
 URL: https://issues.apache.org/jira/browse/LUCENE-6504
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Robert Muir
 Attachments: LUCENE-6504.patch


 We added this api in LUCENE-5729 but we never explored implementing norms 
 with it. These are generally the largest consumer of heap memory and often a 
 real hassle for users.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-7600) Many query functions will cast their operands in float and loose precision in the process.

2015-05-27 Thread Daniel Shane (JIRA)
Daniel Shane created SOLR-7600:
--

 Summary: Many query functions will cast their operands in float 
and loose precision in the process.
 Key: SOLR-7600
 URL: https://issues.apache.org/jira/browse/SOLR-7600
 Project: Solr
  Issue Type: Bug
  Components: Server
Affects Versions: 4.10.3
Reporter: Daniel Shane
Priority: Minor


Function Queries like the 'max' function will cast its result to a float value 
even if the source value cannot be converted to floating point without loosing 
precision.

For example, the max() function will convert dates to floats, and in the 
process we loose some precision (milliseconds). This is problematic if we want 
to sort afterwards since we do not have a millisecond precision anymore. I do 
not know if there is a work around short of creating a new set of query 
functions that would take longs / dates / etc... and return the corresponding 
type and name them 'long_max(), date_max() etc...'

I believe it would be more intuitive if functions like max() would return the 
same type as what they got in their argument (assuming they are all of the same 
type). 

max(date, date) should return a date
max(long, long) should return a long



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Solr-Artifacts-5.x - Build # 844 - Failure

2015-05-27 Thread Mark Miller
Hmm...this is odd. Fails on my local jenkins machine as well but not on my
dev machine (using Java 7 JVM).

- Mark

On Wed, May 27, 2015 at 11:01 AM Apache Jenkins Server 
jenk...@builds.apache.org wrote:

 Build: https://builds.apache.org/job/Solr-Artifacts-5.x/844/

 No tests ran.

 Build Log:
 [...truncated 12816 lines...]
 [javac] Compiling 807 source files to
 /x1/jenkins/jenkins-slave/workspace/Solr-Artifacts-5.x/solr/build/solr-core/classes/java
 [javac]
 /x1/jenkins/jenkins-slave/workspace/Solr-Artifacts-5.x/solr/core/src/java/org/apache/solr/logging/MDCLoggingContext.java:26:
 error: package java.util.function does not exist
 [javac] import java.util.function.Supplier;
 [javac]  ^
 [javac]
 /x1/jenkins/jenkins-slave/workspace/Solr-Artifacts-5.x/solr/core/src/java/org/apache/solr/logging/MDCLoggingContext.java:42:
 error: cannot find symbol
 [javac]   private static ThreadLocalInteger CALL_DEPTH =
 ThreadLocal.withInitial(new SupplierInteger() {
 [javac]
 ^
 [javac]   symbol:   class Supplier
 [javac]   location: class MDCLoggingContext
 [javac] Note: Some input files use or override a deprecated API.
 [javac] Note: Recompile with -Xlint:deprecation for details.
 [javac] Note: Some input files use unchecked or unsafe operations.
 [javac] Note: Recompile with -Xlint:unchecked for details.
 [javac] 2 errors

 BUILD FAILED
 /x1/jenkins/jenkins-slave/workspace/Solr-Artifacts-5.x/solr/build.xml:510:
 The following error occurred while executing this line:
 /x1/jenkins/jenkins-slave/workspace/Solr-Artifacts-5.x/solr/common-build.xml:389:
 The following error occurred while executing this line:
 /x1/jenkins/jenkins-slave/workspace/Solr-Artifacts-5.x/lucene/common-build.xml:523:
 The following error occurred while executing this line:
 /x1/jenkins/jenkins-slave/workspace/Solr-Artifacts-5.x/lucene/common-build.xml:1945:
 Compile failed; see the compiler error output for details.

 Total time: 1 minute 26 seconds
 Build step 'Invoke Ant' marked build as failure
 Archiving artifacts
 Sending artifact delta relative to Solr-Artifacts-5.x #843
 Archived 3 artifacts
 Archive block size is 32768
 Received 0 blocks and 38101191 bytes
 Compression is 0.0%
 Took 1 min 12 sec
 Publishing Javadoc
 Email was triggered for: Failure
 Sending email for trigger: Failure



 -
 To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
 For additional commands, e-mail: dev-h...@lucene.apache.org


[jira] [Commented] (LUCENE-6481) Improve GeoPointField type to only visit high precision boundary terms

2015-05-27 Thread Michael McCandless (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6481?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561052#comment-14561052
 ] 

Michael McCandless commented on LUCENE-6481:


I fixed the randomized test case to only test a smallish global bbox (up to 
+/- 1.5 in lat and lon directions)... when testing the full space the test runs 
very very slowly, even testRandomTiny, because it can require 100s of K terms 
... I'm not quite sure why (?) but the query classes do advertise that they 
should be used on smallish rectangles.

Also, the test fails because of boundary cases:
{noformat}
T0: iter=57 id=6819 docID=6723 lat=-81.12143028547105 lon=-168.98618510239544 
(bbox: lat=-81.22948018485512 TO -80.9313958433031 lon=-168.98618505380117 TO 
-168.77958782241174) expected true but got: false deleted?=false 
query=GeoPointInBBoxQuery: field=geoField: Lower Left: 
[-168.98618505380117,-81.22948018485512] Upper Right: 
[-168.77958782241174,-80.9313958433031]
máj 27, 2015 2:16:52 PM 
com.carrotsearch.randomizedtesting.RandomizedRunner$QueueUncaughtExceptionsHandler
 uncaughtException
WARNING: Uncaught exception in thread: Thread[T0,5,TGRP-TestGeoPointQuery]
java.lang.AssertionError: wrong result
at __randomizedtesting.SeedInfo.seed([4B5245DED09AE592]:0)
at org.junit.Assert.fail(Assert.java:93)
at 
org.apache.lucene.search.TestGeoPointQuery$1._run(TestGeoPointQuery.java:380)
at 
org.apache.lucene.search.TestGeoPointQuery$1.run(TestGeoPointQuery.java:279)
{noformat}

It's odd because in GeoUtils#bboxContains, we do try to take the quantization 
into account ...

 Improve GeoPointField type to only visit high precision boundary terms 
 ---

 Key: LUCENE-6481
 URL: https://issues.apache.org/jira/browse/LUCENE-6481
 Project: Lucene - Core
  Issue Type: Improvement
  Components: core/index
Reporter: Nicholas Knize
 Attachments: LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481.patch, 
 LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481.patch, LUCENE-6481_WIP.patch


 Current GeoPointField [LUCENE-6450 | 
 https://issues.apache.org/jira/browse/LUCENE-6450] computes a set of ranges 
 along the space-filling curve that represent a provided bounding box.  This 
 determines which terms to visit in the terms dictionary and which to skip. 
 This is suboptimal for large bounding boxes as we may end up visiting all 
 terms (which could be quite large). 
 This incremental improvement is to improve GeoPointField to only visit high 
 precision terms in boundary ranges and use the postings list for ranges that 
 are completely within the target bounding box.
 A separate improvement is to switch over to auto-prefix and build an 
 Automaton representing the bounding box.  That can be tracked in a separate 
 issue.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-7115) UpdateLog can miss closing transaction log objects.

2015-05-27 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7115?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561238#comment-14561238
 ] 

Mark Miller commented on SOLR-7115:
---

It's possible this was caused by SOLR-7478.

 UpdateLog can miss closing transaction log objects.
 ---

 Key: SOLR-7115
 URL: https://issues.apache.org/jira/browse/SOLR-7115
 Project: Solr
  Issue Type: Bug
Reporter: Mark Miller

 I've seen this happen on YourKit and in various tests - especially since 
 adding resource release tracking to the log objects. Now I've got a test that 
 catches it in SOLR-7113.
 It seems that in precommit, if prevTlog is not null, we need to close it 
 because we are going to overwrite prevTlog with a new log.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-7601) If a test fails, that error should be reported and not an error about resources that were not closed later.

2015-05-27 Thread Dawid Weiss (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7601?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561331#comment-14561331
 ] 

Dawid Weiss commented on SOLR-7601:
---

This should be close. AfterClass is invoked inside the class rule that detects 
failures so certain exceptions could happen in between, but individual tests 
should taint that suiteFailureMarker.

You will not be able to react to hanging threads -- this is initiated and 
executed outside of the test classes, followed by interrupt on the test thread 
(which then may throw random crap at the fan). This is chicken-and-egg problem, 
I don't think it's quite solvable. You should scan for failures in the log, 
top-to-bottom...

 If a test fails, that error should be reported and not an error about 
 resources that were not closed later.
 ---

 Key: SOLR-7601
 URL: https://issues.apache.org/jira/browse/SOLR-7601
 Project: Solr
  Issue Type: Test
Reporter: Mark Miller
Assignee: Mark Miller
 Fix For: Trunk, 5.3

 Attachments: SOLR-7601.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Solr Ref Guide 5.2 RC Planning - aiming for RC0 on 2015-06-01

2015-05-27 Thread Chris Hostetter


I'm volunteering to be the RM for the 5.2 ref guide.

Now that anshum has started on the 5.2 (code) RC process, we need to 
knuckle down on the ref guide as well.  I'm about to start reviewing  
summarizing the CHANGES.txt list for notable things that need doc'ed which 
i'll post on the TODO page -- but in the meantime if you've worked on a 
new Solr feature in 5.2, please make sure you've updated the corrisponding 
page in the ref guide -- you shouldn't be waiting until the last minute to 
do that, this part of the process is just about looking for things that 
have fallen through the cracks.


My goal is to cut the first ref guide RC no later then monday (2015-06-01) 
and -- as usual -- cancel any ref guide vote that is ongoing if/when a 
new code RC is published tha affects anything in the ref guide.




-Hoss
http://www.lucidworks.com/

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: Hmm, bug?

2015-05-27 Thread Erick Erickson
Problem here is that you changed the schema without completely
re-indexing. In fact, when changes like this are made I typically blow
away the entire data directory or delete/recreate the collection.

At the Lucene layer, there's no schema and certainly no going back to
update already closed segments because a schema changed. So things get
confused when trying to sort since multiple segments have different
definitions for the same field. The loose coupling between the schema
and what's already in the index from an older schema definition is one
of the things we learn to deal with.

In general, when doing much of anything to the schema except _adding_
fields, I recommend starting over completely. You can do this crudely
by stopping all your servers and deleting the data directory (as in
'rm -rf data) then restarting. Or, just use the collections API to
delete then re-add the collection.

Best,
Erick

On Wed, May 27, 2015 at 9:21 AM, Gus Heck gus.h...@gmail.com wrote:
 So I did the following (seemingly) reasonable set of steps...

 Set up a Solr Cloud instance with mutable managed schema (data driven
 config)
 Tossed a couple docs in just to see it work.
 Tried to sort on 'title', realized title had been guessed as multivalued.
 Changed the field to type=string (as opposed to strings) using schema api
 Tried to sort on title

 The result is

 trace: java.lang.IllegalStateException: unexpected docvalues type
 SORTED_SET for field 'title' (expected=SORTED). Use UninvertingReader or
 index with docvalues.\n\tat
 org.apache.lucene.index.DocValues.checkField(DocValues.java:208)

 Resending the curl command to re-index the docs hangs

 However, I have not played with schemaless or the schema api before, so
 maybe I've missed something, and my steps are not reasonable, and thus this
 sanity check before I file a bug.

 -Gus

 --
 http://www.the111shift.com

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-7601) If a test fails, that error should be reported and not an error about resources that were not closed later.

2015-05-27 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7601?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561280#comment-14561280
 ] 

Mark Miller commented on SOLR-7601:
---

Hmm...that is not quite what I'm after. Another attempt coming.

Trying to solve test failures where jenkins reports things like searchers not 
being closed, but the real issue is something else (like hanging threads or 
something else).

 If a test fails, that error should be reported and not an error about 
 resources that were not closed later.
 ---

 Key: SOLR-7601
 URL: https://issues.apache.org/jira/browse/SOLR-7601
 Project: Solr
  Issue Type: Test
Reporter: Mark Miller
Assignee: Mark Miller
 Fix For: Trunk, 5.3

 Attachments: SOLR-7601.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-7601) If a test fails, that error should be reported and not an error about resources that were not closed later.

2015-05-27 Thread Mark Miller (JIRA)
Mark Miller created SOLR-7601:
-

 Summary: If a test fails, that error should be reported and not an 
error about resources that were not closed later.
 Key: SOLR-7601
 URL: https://issues.apache.org/jira/browse/SOLR-7601
 Project: Solr
  Issue Type: Test
Reporter: Mark Miller
Assignee: Mark Miller
 Fix For: Trunk, 5.3
 Attachments: SOLR-7601.patch





--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-7601) If a test fails, that error should be reported and not an error about resources that were not closed later.

2015-05-27 Thread Mark Miller (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-7601?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Miller updated SOLR-7601:
--
Attachment: SOLR-7601.patch

 If a test fails, that error should be reported and not an error about 
 resources that were not closed later.
 ---

 Key: SOLR-7601
 URL: https://issues.apache.org/jira/browse/SOLR-7601
 Project: Solr
  Issue Type: Test
Reporter: Mark Miller
Assignee: Mark Miller
 Fix For: Trunk, 5.3

 Attachments: SOLR-7601.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: Hmm, bug (schemaless/schema api)?

2015-05-27 Thread Gus Heck
Ok, thanks.

Sorry for the uninformative subject line. I realized after the fact it was
not very good (ok really bad :) ).

What you say is what I had generally understood to be the case in the past
with regular static schemas, but I had hoped something newer and more
flexible was implied with the advent of an API, and the schemaless
configuration. Although there is a caveat at the top of the wiki page,
statements like

When modifying the schema with the REST API, a core reload will
 automatically occur in order for the changes to be available immediately.


had given me hope. Guess you haven't got that far after all.

-Gus


On Wed, May 27, 2015 at 1:03 PM, Erick Erickson erickerick...@gmail.com
wrote:

 Problem here is that you changed the schema without completely
 re-indexing. In fact, when changes like this are made I typically blow
 away the entire data directory or delete/recreate the collection.

 At the Lucene layer, there's no schema and certainly no going back to
 update already closed segments because a schema changed. So things get
 confused when trying to sort since multiple segments have different
 definitions for the same field. The loose coupling between the schema
 and what's already in the index from an older schema definition is one
 of the things we learn to deal with.

 In general, when doing much of anything to the schema except _adding_
 fields, I recommend starting over completely. You can do this crudely
 by stopping all your servers and deleting the data directory (as in
 'rm -rf data) then restarting. Or, just use the collections API to
 delete then re-add the collection.

 Best,
 Erick

 On Wed, May 27, 2015 at 9:21 AM, Gus Heck gus.h...@gmail.com wrote:
  So I did the following (seemingly) reasonable set of steps...
 
  Set up a Solr Cloud instance with mutable managed schema (data driven
  config)
  Tossed a couple docs in just to see it work.
  Tried to sort on 'title', realized title had been guessed as multivalued.
  Changed the field to type=string (as opposed to strings) using schema api
  Tried to sort on title
 
  The result is
 
  trace: java.lang.IllegalStateException: unexpected docvalues type
  SORTED_SET for field 'title' (expected=SORTED). Use UninvertingReader or
  index with docvalues.\n\tat
  org.apache.lucene.index.DocValues.checkField(DocValues.java:208)
 
  Resending the curl command to re-index the docs hangs
 
  However, I have not played with schemaless or the schema api before, so
  maybe I've missed something, and my steps are not reasonable, and thus
 this
  sanity check before I file a bug.
 
  -Gus
 
  --
  http://www.the111shift.com

 -
 To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
 For additional commands, e-mail: dev-h...@lucene.apache.org




-- 
http://www.the111shift.com


[jira] [Commented] (SOLR-7118) ChaosMonkeyNothingIsSafeTest fails with too many update fails

2015-05-27 Thread Mark Miller (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7118?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561241#comment-14561241
 ] 

Mark Miller commented on SOLR-7118:
---

Saw another errant fail on jenkins cluster - much, much rarer at least - going 
to raise again.

 ChaosMonkeyNothingIsSafeTest fails with too many update fails
 -

 Key: SOLR-7118
 URL: https://issues.apache.org/jira/browse/SOLR-7118
 Project: Solr
  Issue Type: Bug
  Components: SolrCloud, Tests
Affects Versions: 5.0
Reporter: Shalin Shekhar Mangar
 Fix For: Trunk, 5.2

 Attachments: SOLR-7118.patch


 There are frequent failures on both trunk and branch_5x with the following 
 message:
 {code}
 java.lang.AssertionError: There were too many update fails - we expect it can 
 happen, but shouldn't easily
   at 
 __randomizedtesting.SeedInfo.seed([786DB0FD42626C16:F98B3EE5353D0C2A]:0)
   at org.junit.Assert.fail(Assert.java:93)
   at org.junit.Assert.assertTrue(Assert.java:43)
   at org.junit.Assert.assertFalse(Assert.java:68)
   at 
 org.apache.solr.cloud.ChaosMonkeyNothingIsSafeTest.doTest(ChaosMonkeyNothingIsSafeTest.java:224)
   at 
 org.apache.solr.BaseDistributedSearchTestCase.testDistribSearch(BaseDistributedSearchTestCase.java:878)
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-7570) Config APIs should not modify the ConfigSet

2015-05-27 Thread Alan Woodward (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-7570?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alan Woodward updated SOLR-7570:

Attachment: SOLR-7570.patch

Here's the first cut of an idea.

* Adds a writeResource(String, byte[], Location) method to SolrResourceLoader 
(Location can be LOCAL, COLLECTION, CONFIGSET)
* SolrResourceLoader takes both a local instance dir and a shared config dir.  
These can be the same if you're not using a configset.
* The standard resource loader looks in three places for resources:
** core instance dir
** configset
** classpath
* The ZK resource loader looks in four places:
** core instance dir
** collection-specific config
** zk config
** classpath

You can write to either the local core instance dir, or to the 
collection-specific config (I added CONFIG as a location in case we want to use 
that later for things like specifying where a particular resource was found, 
but that can be taken out if it's not adding anything now).

Writing to the collection-specific config uses version-tracking to implement 
optimistic concurrency.

There are tests for the standard resource loader and the ZK resource loader.

This is still pretty rough around the edges, and I haven't run the full test 
suite or started cutting over existing code to using the new API, but it's a 
start.  What do people think?

 Config APIs should not modify the ConfigSet
 ---

 Key: SOLR-7570
 URL: https://issues.apache.org/jira/browse/SOLR-7570
 Project: Solr
  Issue Type: Improvement
Reporter: Tomás Fernández Löbbe
 Attachments: SOLR-7570.patch


 Originally discussed here: 
 http://mail-archives.apache.org/mod_mbox/lucene-dev/201505.mbox/%3CCAMJgJxSXCHxDzJs5-C-pKFDEBQD6JbgxB=-xp7u143ekmgp...@mail.gmail.com%3E
 The ConfigSet used to create a collection should be read-only. Changes made 
 via any of the Config APIs should only be applied to the collection where the 
 operation is done and no to other collections that may be using the same 
 ConfigSet. As discussed in the dev list: 
 When a collection is created we should have two things, an immutable part 
 (the ConfigSet) and a mutable part (configoverlay, generated schema, etc). 
 The ConfigSet will still be placed in ZooKeeper under /configs but the 
 mutable part should be placed under /collections/$COLLECTION_NAME/…
 [~romseygeek] suggested: 
 {quote}
 A nice way of doing it would be to make it part of the SolrResourceLoader 
 interface.  The ZK resource loader could check in the collection-specific 
 zknode first, and then under configs/, and we could add a writeResource() 
 method that writes to the collection-specific node as well.  Then all config 
 I/O goes via the resource loader, and we have a way of keeping certain parts 
 immutable.
 {quote}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-7344) Use two thread pools, one for internal requests and one for external, to avoid distributed deadlock and decrease the number of threads that need to be created.

2015-05-27 Thread Yonik Seeley (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7344?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561627#comment-14561627
 ] 

Yonik Seeley commented on SOLR-7344:


bq. Note even for this approach, we need to implement a way to identify an 
internal request

Shard sub-requests will always have isShard=true param set.

bq.  The idea here is to implement a custom Servlet filter similar to Jetty 
QoSFilter

Looks interesting. I wonder what the performance implications of using 
continuations are?

 Use two thread pools, one for internal requests and one for external, to 
 avoid distributed deadlock and decrease the number of threads that need to be 
 created.
 ---

 Key: SOLR-7344
 URL: https://issues.apache.org/jira/browse/SOLR-7344
 Project: Solr
  Issue Type: Improvement
  Components: SolrCloud
Reporter: Mark Miller
 Attachments: SOLR-7344.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-7602) Frequent MultiThreadedOCPTest failures on Jenkins

2015-05-27 Thread Anshum Gupta (JIRA)
Anshum Gupta created SOLR-7602:
--

 Summary: Frequent MultiThreadedOCPTest failures on Jenkins
 Key: SOLR-7602
 URL: https://issues.apache.org/jira/browse/SOLR-7602
 Project: Solr
  Issue Type: Bug
Reporter: Anshum Gupta


The number of failed MultiThreadedOCPTest runs on Jenkins has gone up 
drastically since Apr 30, 2015.
{code}
REGRESSION:  org.apache.solr.cloud.MultiThreadedOCPTest.test

Error Message:
Captured an uncaught exception in thread: Thread[id=6313, 
name=parallelCoreAdminExecutor-1988-thread-15, state=RUNNABLE, 
group=TGRP-MultiThreadedOCPTest]

Stack Trace:
com.carrotsearch.randomizedtesting.UncaughtExceptionError: Captured an uncaught 
exception in thread: Thread[id=6313, 
name=parallelCoreAdminExecutor-1988-thread-15, state=RUNNABLE, 
group=TGRP-MultiThreadedOCPTest]
at 
__randomizedtesting.SeedInfo.seed([1FD11A82D96D185B:97852558779175A3]:0)
Caused by: java.lang.AssertionError: Too many closes on SolrCore
at __randomizedtesting.SeedInfo.seed([1FD11A82D96D185B]:0)
at org.apache.solr.core.SolrCore.close(SolrCore.java:1138)
at org.apache.solr.common.util.IOUtils.closeQuietly(IOUtils.java:31)
at org.apache.solr.core.CoreContainer.create(CoreContainer.java:535)
at org.apache.solr.core.CoreContainer.create(CoreContainer.java:494)
at 
org.apache.solr.handler.admin.CoreAdminHandler.handleCreateAction(CoreAdminHandler.java:598)
at 
org.apache.solr.handler.admin.CoreAdminHandler.handleRequestInternal(CoreAdminHandler.java:212)
at 
org.apache.solr.handler.admin.CoreAdminHandler$ParallelCoreAdminHandlerThread.run(CoreAdminHandler.java:1219)
at 
org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor$1.run(ExecutorUtil.java:148)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
{code}

Last failure:
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-5.x-Linux/12665/



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-6485) Add a custom separator break iterator

2015-05-27 Thread Robert Muir (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-6485?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Muir resolved LUCENE-6485.
-
   Resolution: Fixed
Fix Version/s: 5.3
   Trunk

Thanks Luca! This is a nice alternative to WholeBreakIterator. 

 Add a custom separator break iterator
 -

 Key: LUCENE-6485
 URL: https://issues.apache.org/jira/browse/LUCENE-6485
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Luca Cavanna
 Fix For: Trunk, 5.3

 Attachments: LUCENE-6485.patch


 Lucene currently includes a WholeBreakIterator used to highlight entire 
 fields using the postings highlighter, without breaking their content into 
 sentences.
 I would like to contribute a CustomSeparatorBreakIterator that breaks when a 
 custom char separator is found in the text. This can be used for instance 
 when wanting to highlight entire fields, value per value. One can subclass 
 PostingsHighlighter and have getMultiValueSeparator return a control 
 character, like U+ , then use the custom break iterator to break on 
 U+ so that one snippet per value will be generated.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-7555) Display total space and available space in Admin

2015-05-27 Thread Eric Pugh (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-7555?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561575#comment-14561575
 ] 

Eric Pugh commented on SOLR-7555:
-

Humm...   You know, I saw that RAMDirectory implemented a interface called 
Accountable that was for tracking memory usage.So we could modify the 
HdfsDirectory and FSDirectory to implement that. 

Then in SystemInfoHandler we do a instanceof check, and then pull that data?

[~ehatcher] [~mariusneo] what do you guys think?

 Display total space and available space in Admin
 

 Key: SOLR-7555
 URL: https://issues.apache.org/jira/browse/SOLR-7555
 Project: Solr
  Issue Type: Improvement
  Components: web gui
Affects Versions: 5.1
Reporter: Eric Pugh
Assignee: Erik Hatcher
Priority: Minor
 Fix For: 5.2

 Attachments: SOLR-7555-display_disk_space.patch, 
 SOLR-7555-display_disk_space_v2.patch, SOLR-7555.patch


 Frequently I have access to the Solr Admin console, but not the underlying 
 server, and I'm curious how much space remains available.   This little patch 
 exposes total Volume size as well as the usable space remaining:
 !https://monosnap.com/file/VqlReekCFwpK6utI3lP18fbPqrGI4b.png!
 I'm not sure if this is the best place to put this, as every shard will share 
 the same data, so maybe it should be on the top level Dashboard?  Also not 
 sure what to call the fields! 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



RE: svn commit: r1682057 - /lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/logging/MDCLoggingContext.java

2015-05-27 Thread Uwe Schindler
Hi Mark,

I am wondering why you be so verbose in Java 8's code. The main reason for 
ThreadLocal.withInitial() is to use it like that:

private static final ThreadLocalInteger CALL_DEPTH = 
ThreadLocal.withInitial(() - 0);

Using a Supplier without a lambda is not the intention behind this method /API 
:-)
In Java 7, you can use the code like this fix comitted. If you want it verbose 
in Java 8, I would also use the Java 7 code...

Uwe

-
Uwe Schindler
H.-H.-Meier-Allee 63, D-28213 Bremen
http://www.thetaphi.de
eMail: u...@thetaphi.de

 -Original Message-
 From: markrmil...@apache.org [mailto:markrmil...@apache.org]
 Sent: Wednesday, May 27, 2015 5:28 PM
 To: comm...@lucene.apache.org
 Subject: svn commit: r1682057 -
 /lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/loggin
 g/MDCLoggingContext.java
 
 Author: markrmiller
 Date: Wed May 27 15:27:43 2015
 New Revision: 1682057
 
 URL: http://svn.apache.org/r1682057
 Log:
 SOLR-7590: Java8 code - Java7 for branch_5x.
 
 Modified:
 
 lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/logging
 /MDCLoggingContext.java
 
 Modified:
 lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/logging
 /MDCLoggingContext.java
 URL:
 http://svn.apache.org/viewvc/lucene/dev/branches/branch_5x/solr/core/sr
 c/java/org/apache/solr/logging/MDCLoggingContext.java?rev=1682057r1=
 1682056r2=1682057view=diff
 ==
 
 ---
 lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/logging
 /MDCLoggingContext.java (original)
 +++
 lucene/dev/branches/branch_5x/solr/core/src/java/org/apache/solr/log
 +++ ging/MDCLoggingContext.java Wed May 27 15:27:43 2015
 @@ -23,7 +23,6 @@ import static org.apache.solr.common.clo  import static
 org.apache.solr.common.cloud.ZkStateReader.REPLICA_PROP;
  import static
 org.apache.solr.common.cloud.ZkStateReader.SHARD_ID_PROP;
 
 -import java.util.function.Supplier;
 
  import org.apache.solr.cloud.CloudDescriptor;
  import org.apache.solr.cloud.ZkController;
 @@ -39,12 +38,13 @@ import org.slf4j.MDC;
   */
  public class MDCLoggingContext {
// When a thread sets context and finds that the context is already set, we
 should noop and ignore the finally clear
 -  private static ThreadLocalInteger CALL_DEPTH =
 ThreadLocal.withInitial(new SupplierInteger() {
 +  private static ThreadLocalInteger CALL_DEPTH = new
 + ThreadLocalInteger() {
  @Override
 -public Integer get() {
 +protected Integer initialValue() {
return 0;
  }
 -  });
 +  };
 +
 
private static void setCollection(String collection) {
  if (collection != null) {



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6459) [suggest] Query Interface for suggest API

2015-05-27 Thread Michael McCandless (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6459?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561616#comment-14561616
 ] 

Michael McCandless commented on LUCENE-6459:


Thanks [~areek], new patch looks great, passes {{ant precommit}} ... I'll 
commit soon.

 [suggest] Query Interface for suggest API
 -

 Key: LUCENE-6459
 URL: https://issues.apache.org/jira/browse/LUCENE-6459
 Project: Lucene - Core
  Issue Type: New Feature
  Components: core/search
Affects Versions: 5.1
Reporter: Areek Zillur
Assignee: Areek Zillur
 Fix For: Trunk, 5.x, 5.1

 Attachments: LUCENE-6459.patch, LUCENE-6459.patch, LUCENE-6459.patch, 
 LUCENE-6459.patch, LUCENE-6459.patch, LUCENE-6459.patch, LUCENE-6459.patch, 
 LUCENE-6459.patch, LUCENE-6459.patch, LUCENE-6459.patch, LUCENE-6459.patch


 This patch factors out common indexing/search API used by the recently 
 introduced [NRTSuggester|https://issues.apache.org/jira/browse/LUCENE-6339]. 
 The motivation is to provide a query interface for FST-based fields 
 (*SuggestField* and *ContextSuggestField*) 
 to enable suggestion scoring and more powerful automaton queries. 
 Previously, only prefix ‘queries’ with index-time weights were supported but 
 we can also support:
 * Prefix queries expressed as regular expressions:  get suggestions that 
 match multiple prefixes
   ** *Example:* _star\[wa\|tr\]_ matches _starwars_ and _startrek_
 * Fuzzy Prefix queries supporting scoring: get typo tolerant suggestions 
 scored by how close they are to the query prefix
 ** *Example:* querying for _seper_ will score _separate_ higher then 
 _superstitious_
 * Context Queries: get suggestions boosted and/or filtered based on their 
 indexed contexts (meta data)
 ** *Boost example:* get typo tolerant suggestions on song names with 
 prefix _like a roling_ boosting songs with 
 genre _rock_ and _indie_
 ** *Filter example:* get suggestion on all file names starting with 
 _finan_ only for _user1_ and _user2_
 h3. Suggest API
 {code}
 SuggestIndexSearcher searcher = new SuggestIndexSearcher(reader);
 CompletionQuery query = ...
 TopSuggestDocs suggest = searcher.suggest(query, num);
 {code}
 h3. CompletionQuery
 *CompletionQuery* is used to query *SuggestField* and *ContextSuggestField*. 
 A *CompletionQuery* produces a *CompletionWeight*, 
 which allows *CompletionQuery* implementations to pass in an automaton that 
 will be intersected with a FST and allows boosting and 
 meta data extraction from the intersected partial paths. A *CompletionWeight* 
 produces a *CompletionScorer*. A *CompletionScorer* 
 executes a Top N search against the FST with the provided automaton, scoring 
 and filtering all matched paths. 
 h4. PrefixCompletionQuery
 Return documents with values that match the prefix of an analyzed term text 
 Documents are sorted according to their suggest field weight. 
 {code}
 PrefixCompletionQuery(Analyzer analyzer, Term term)
 {code}
 h4. RegexCompletionQuery
 Return documents with values that match the prefix of a regular expression
 Documents are sorted according to their suggest field weight.
 {code}
 RegexCompletionQuery(Term term)
 {code}
 h4. FuzzyCompletionQuery
 Return documents with values that has prefixes within a specified edit 
 distance of an analyzed term text.
 Documents are ‘boosted’ by the number of matching prefix letters of the 
 suggestion with respect to the original term text.
 {code}
 FuzzyCompletionQuery(Analyzer analyzer, Term term)
 {code}
 h5. Scoring
 {{suggestion_weight * boost}}
 where {{suggestion_weight}} and {{boost}} are all integers. 
 {{boost = # of prefix characters matched}}
 h4. ContextQuery
 Return documents that match a {{CompletionQuery}} filtered and/or boosted by 
 provided context(s). 
 {code}
 ContextQuery(CompletionQuery query)
 contextQuery.addContext(CharSequence context, int boost, boolean exact)
 {code}
 *NOTE:* {{ContextQuery}} should be used with {{ContextSuggestField}} to query 
 suggestions boosted and/or filtered by contexts.
 Running {{ContextQuery}} against a {{SuggestField}} will error out.
 h5. Scoring
 {{suggestion_weight  * context_boost}}
 where {{suggestion_weight}} and {{context_boost}} are all integers
 When used with {{FuzzyCompletionQuery}},
 {{suggestion_weight * (context_boost + fuzzy_boost)}}
 h3. Context Suggest Field
 To use {{ContextQuery}}, use {{ContextSuggestField}} instead of 
 {{SuggestField}}. Any {{CompletionQuery}} can be used with 
 {{ContextSuggestField}}, the default behaviour is to return suggestions from 
 *all* contexts. {{Context}} for every completion hit 
 can be accessed through {{SuggestScoreDoc#context}}.
 {code}
 ContextSuggestField(String name, CollectionCharSequence contexts, String 
 value, int weight) 
 {code}



--
This 

[VOTE] 5.2.0 RC1

2015-05-27 Thread Anshum Gupta
Please vote for the first release candidate for Lucene/Solr 5.2.0

The artifacts can be downloaded from:
https://dist.apache.org/repos/dist/dev/lucene/lucene-solr-5.2.0-RC1-rev1682085

You can run the smoke tester directly with this command:

python3 -u dev-tools/scripts/smokeTestRelease.py
https://dist.apache.org/repos/dist/dev/lucene/lucene-solr-5.2.0-RC1-rev1682085

I've run the test suite a few times, the smoke tester, basic collection
creation, startup, indexing, and query. Here's my +1.

SUCCESS! [0:31:23.943785]

P.S: I hit failure in MultiThreadedOCPTest 2 times while creating the RC,
so I'm looking at what's triggering it in parallel to make sure that we're
not overlooking a problem. As it has been failing on Jenkins frequently,
I've created SOLR-7602 https://issues.apache.org/jira/browse/SOLR-7602to
track this.

--
Anshum Gupta


[jira] [Commented] (LUCENE-6485) Add a custom separator break iterator

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6485?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561622#comment-14561622
 ] 

ASF subversion and git services commented on LUCENE-6485:
-

Commit 1682114 from [~rcmuir] in branch 'dev/trunk'
[ https://svn.apache.org/r1682114 ]

LUCENE-6485: Add CustomSeparatorBreakIterator

 Add a custom separator break iterator
 -

 Key: LUCENE-6485
 URL: https://issues.apache.org/jira/browse/LUCENE-6485
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Luca Cavanna
 Attachments: LUCENE-6485.patch


 Lucene currently includes a WholeBreakIterator used to highlight entire 
 fields using the postings highlighter, without breaking their content into 
 sentences.
 I would like to contribute a CustomSeparatorBreakIterator that breaks when a 
 custom char separator is found in the text. This can be used for instance 
 when wanting to highlight entire fields, value per value. One can subclass 
 PostingsHighlighter and have getMultiValueSeparator return a control 
 character, like U+ , then use the custom break iterator to break on 
 U+ so that one snippet per value will be generated.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-6485) Add a custom separator break iterator

2015-05-27 Thread ASF subversion and git services (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-6485?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14561638#comment-14561638
 ] 

ASF subversion and git services commented on LUCENE-6485:
-

Commit 1682115 from [~rcmuir] in branch 'dev/branches/branch_5x'
[ https://svn.apache.org/r1682115 ]

LUCENE-6485: Add CustomSeparatorBreakIterator

 Add a custom separator break iterator
 -

 Key: LUCENE-6485
 URL: https://issues.apache.org/jira/browse/LUCENE-6485
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Luca Cavanna
 Attachments: LUCENE-6485.patch


 Lucene currently includes a WholeBreakIterator used to highlight entire 
 fields using the postings highlighter, without breaking their content into 
 sentences.
 I would like to contribute a CustomSeparatorBreakIterator that breaks when a 
 custom char separator is found in the text. This can be used for instance 
 when wanting to highlight entire fields, value per value. One can subclass 
 PostingsHighlighter and have getMultiValueSeparator return a control 
 character, like U+ , then use the custom break iterator to break on 
 U+ so that one snippet per value will be generated.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



  1   2   >