RE: verbose logging

2013-03-04 Thread Uwe Schindler
Yeah,

it would be nice to have those, because with the Jenkins Web GUI you could then 
inspect stdout from each test, like that one from forbiddenAPIs:
http://jenkins.thetaphi.de/job/Forbidden-APIs/119/testReport/src.test.antunit/TestClassDeclaration_xml/testForbiddenExtends/?
Can we maybe limit the amount and only have something like "tail -", so you 
get at least the recent stdout.
I know that the plain text files are easier to write/manage, but the XML files 
are good for inspection of test output with Jenkins. Currently the XML output 
is only filled when test fail.

Uwe

-
Uwe Schindler
H.-H.-Meier-Allee 63, D-28213 Bremen
http://www.thetaphi.de
eMail: u...@thetaphi.de


> -Original Message-
> From: dawid.we...@gmail.com [mailto:dawid.we...@gmail.com] On Behalf
> Of Dawid Weiss
> Sent: Tuesday, March 05, 2013 8:00 AM
> To: dev@lucene.apache.org
> Subject: Re: verbose logging
> 
> common-build.xml:
> 
> 
>  outputStreams="no" />
> 
> Like I said though -- I would *not* bring stdout/stderr back to those XML 
> files
> unless you want to deal with other problems resulting from
> 10+mb XML files. They also take a lot longer to serialize/ save than
> just plain text output.
> 
> Dawid
> 
> 
> On Tue, Mar 5, 2013 at 7:58 AM, Dawid Weiss
>  wrote:
> >> I though the XML report files contain the whole stdout?
> >>
> >> http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/ws/solr/build/so
> >> lr-core/test/TEST-org.apache.solr.update.TestAtomicUpdateErrorCases.x
> >> ml
> >>
> >> But its empty... This would be the most "correct" variant. I know that by
> default stock JUnit saves the whole stdout/stderr in the XML file, but JUnit4
> runner does not seem to do this?
> >
> > I used to emit stdout to those XML files but Solr emits so much
> > logging information from certain tests that it got XML parsers/
> > transformers hit OOM a few times on jenkins... I disabled it
> > intentionally. Those plain-text files are still saved and they're more
> > compact than XML output. I kept the XML output so that jenkins has
> > stuff for its own analysis pipeline but standard output is not needed
> > there.
> >
> > Dawid
> 
> -
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional
> commands, e-mail: dev-h...@lucene.apache.org


-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: verbose logging

2013-03-04 Thread Dawid Weiss
common-build.xml:




Like I said though -- I would *not* bring stdout/stderr back to those
XML files unless you want to deal with other problems resulting from
10+mb XML files. They also take a lot longer to serialize/ save than
just plain text output.

Dawid


On Tue, Mar 5, 2013 at 7:58 AM, Dawid Weiss
 wrote:
>> I though the XML report files contain the whole stdout?
>>
>> http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/ws/solr/build/solr-core/test/TEST-org.apache.solr.update.TestAtomicUpdateErrorCases.xml
>>
>> But its empty... This would be the most "correct" variant. I know that by 
>> default stock JUnit saves the whole stdout/stderr in the XML file, but 
>> JUnit4 runner does not seem to do this?
>
> I used to emit stdout to those XML files but Solr emits so much
> logging information from certain tests that it got XML parsers/
> transformers hit OOM a few times on jenkins... I disabled it
> intentionally. Those plain-text files are still saved and they're more
> compact than XML output. I kept the XML output so that jenkins has
> stuff for its own analysis pipeline but standard output is not needed
> there.
>
> Dawid

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: verbose logging

2013-03-04 Thread Dawid Weiss
> I though the XML report files contain the whole stdout?
>
> http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/ws/solr/build/solr-core/test/TEST-org.apache.solr.update.TestAtomicUpdateErrorCases.xml
>
> But its empty... This would be the most "correct" variant. I know that by 
> default stock JUnit saves the whole stdout/stderr in the XML file, but JUnit4 
> runner does not seem to do this?

I used to emit stdout to those XML files but Solr emits so much
logging information from certain tests that it got XML parsers/
transformers hit OOM a few times on jenkins... I disabled it
intentionally. Those plain-text files are still saved and they're more
compact than XML output. I kept the XML output so that jenkins has
stuff for its own analysis pipeline but standard output is not needed
there.

Dawid

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-4503) Add REST API methods to get schema information: fields, dynamicFields, fieldTypes, and copyFields

2013-03-04 Thread Steve Rowe (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4503?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Rowe updated SOLR-4503:
-

Attachment: SOLR-4503.patch

{quote}
Also related: SOLR-4210
We should aim for being able to hit any node in the cluster w/o worrying about 
which nodes are hosting which collections.
{quote}

I set up a cluster with two collections with different configurations so I 
could easily verify I was getting responses from the right collection.  Turns 
out that the schema info GET requests were being converted into POST requests 
as a side-effect of calling {{con.getOutputStream()}} in 
{{SolrDispatchFilter.remoteQuery()}}, and failing as a result.  In the patch 
snippet below, which is included in the attached patch, I skip forwarding the 
request body unless the original request's method is POST: 

{code:java}
@@ -353,13 +365,17 @@
   try {
 con.connect();
 
-InputStream is = req.getInputStream();
-OutputStream os = con.getOutputStream();
-try {
-  IOUtils.copyLarge(is, os);
-} finally {
-  IOUtils.closeQuietly(os);
-  IOUtils.closeQuietly(is);  // TODO: I thought we weren't supposed to 
explicitly close servlet streams
+InputStream is;
+OutputStream os;
+if ("POST".equals(req.getMethod())) {
+  is = req.getInputStream();
+  os = con.getOutputStream(); // side effect: method is switched to 
POST
+  try {
+IOUtils.copyLarge(is, os);
+  } finally {
+IOUtils.closeQuietly(os);
+IOUtils.closeQuietly(is);  // TODO: I thought we weren't supposed 
to explicitly close servlet streams
+  }
 }
 
 resp.setStatus(con.getResponseCode());
{code}

The patch also contains some fixed tests for the schema REST API (I didn't 
update the tests after I removed printing of copyField maxChars when it's zero).

'ant test' under Solr passes, as does 'ant precommit'.

If there are no objections, I'll commit this form of the patch in a day or so.

> Add REST API methods to get schema information: fields, dynamicFields, 
> fieldTypes, and copyFields
> -
>
> Key: SOLR-4503
> URL: https://issues.apache.org/jira/browse/SOLR-4503
> Project: Solr
>  Issue Type: Sub-task
>  Components: Schema and Analysis
>Affects Versions: 4.1
>Reporter: Steve Rowe
>Assignee: Steve Rowe
> Fix For: 4.2
>
> Attachments: all.dynamic.fields.json, all.dynamic.fields.json, 
> all.fields.json, all.fields.json, all.field.types.json, all.field.types.json, 
> coordinate.dynamic.field.json, coordinate.dynamic.field.json, 
> copyfields.json, date.field.type.json, date.field.type.json, 
> price.field.json, price.field.json, SOLR-4503.patch, SOLR-4503.patch, 
> SOLR-4503.patch, SOLR-4503.patch
>
>
> Add REST methods that provide properties for fields, dynamicFields, 
> fieldTypes, and copyFields, using paths:
> /solr/(corename)/schema/fields
> /solr/(corename)/schema/fields/fieldname
> /solr/(corename)/schema/dynamicfields
> /solr/(corename)/schema/dynamicfields/pattern
> /solr/(corename)/schema/fieldtypes
> /solr/(corename)/schema/fieldtypes/typename 
> /solr/(corename)/schema/copyfields

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-3239) DocTransformer for formatting CurrencyField as a virtual field

2013-03-04 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-3239?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man updated SOLR-3239:
---

Summary: DocTransformer for formatting CurrencyField as a virtual field  
(was: Function for returning a CurrencyField as a virtual field)

FYI: SOLR-4138 added a currency(...) function like the one i proposed.  the 
only thing left is a DocTransformer for arbitrary currency formatting.

> DocTransformer for formatting CurrencyField as a virtual field
> --
>
> Key: SOLR-3239
> URL: https://issues.apache.org/jira/browse/SOLR-3239
> Project: Solr
>  Issue Type: New Feature
>Reporter: Jan Høydahl
> Fix For: 4.2, 5.0
>
>
> With the new CurrencyField from SOLR-2202, we can now do range queries, sort 
> and soon faceting on CurrencyField.
> One thing we cannot do is return a converted value. We can only return the 
> stored value which must be in the same currency as it was entered, e.g. 
> "10.0,USD". People now have to do this conversion in application layer, 
> reading the same currency.xml file as in Solr. This is a mess to keep in 
> sync, and is hard to support with other ExchangeRateProviders or in other 
> client languages than Java.
> This patch should implement a Function which takes a field name and requested 
> currency as input and returns the converted value. Optionally it could also 
> take care of localized display, e.g. "$10.0" instead of "10.0,USD". Proposed 
> syntax:
> {code}
> &fl=title,price:currency(price_c,EUR) # Returns price_c in EUR as 
> "price"
> &fl=title,price:currency(price_c,EUR,en_US)   # Returns price_c in EUR, 
> formatted according to en_US locale
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-Tests-4.x-java7 - Build # 1040 - Failure

2013-03-04 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-Tests-4.x-java7/1040/

All tests passed

Build Log:
[...truncated 21322 lines...]
check-licenses:
 [echo] License check under: 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-java7/solr
 [licenses] MISSING sha1 checksum file for: 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-java7/solr/contrib/extraction/lib/fontbox-1.7.1.jar
 [licenses] MISSING sha1 checksum file for: 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-java7/solr/contrib/extraction/lib/jempbox-1.7.1.jar

 [licenses] MISSING sha1 checksum file for: 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-java7/solr/contrib/extraction/lib/metadata-extractor-2.6.2.jar
 [licenses] MISSING sha1 checksum file for: 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-java7/solr/contrib/extraction/lib/pdfbox-1.7.1.jar
 [licenses] MISSING sha1 checksum file for: 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-java7/solr/contrib/extraction/lib/tika-core-1.3.jar
 [licenses] MISSING sha1 checksum file for: 
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-java7/solr/contrib/extraction/lib/tika-parsers-1.3.jar

[...truncated 1 lines...]
BUILD FAILED
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-java7/build.xml:381:
 The following error occurred while executing this line:
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-java7/build.xml:67:
 The following error occurred while executing this line:
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-java7/solr/build.xml:234:
 The following error occurred while executing this line:
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-Tests-4.x-java7/lucene/tools/custom-tasks.xml:43:
 License check failed. Check the logs.

Total time: 78 minutes 23 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Recording test results
Email was triggered for: Failure
Sending email for trigger: Failure



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[JENKINS] Lucene-Solr-4.x-Linux (32bit/jdk1.7.0_15) - Build # 4543 - Still Failing!

2013-03-04 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/4543/
Java: 32bit/jdk1.7.0_15 -server -XX:+UseParallelGC

All tests passed

Build Log:
[...truncated 21428 lines...]
check-licenses:
 [echo] License check under: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/contrib/extraction/lib/fontbox-1.7.1.jar
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/contrib/extraction/lib/jempbox-1.7.1.jar

 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/contrib/extraction/lib/metadata-extractor-2.6.2.jar
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/contrib/extraction/lib/pdfbox-1.7.1.jar
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/contrib/extraction/lib/tika-core-1.3.jar
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/contrib/extraction/lib/tika-parsers-1.3.jar

[...truncated 1 lines...]
BUILD FAILED
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/build.xml:381: The following 
error occurred while executing this line:
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/build.xml:67: The following 
error occurred while executing this line:
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/build.xml:234: The 
following error occurred while executing this line:
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/tools/custom-tasks.xml:43:
 License check failed. Check the logs.

Total time: 42 minutes 55 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Recording test results
Description set: Java: 32bit/jdk1.7.0_15 -server -XX:+UseParallelGC
Email was triggered for: Failure
Sending email for trigger: Failure



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Commented] (SOLR-4416) Upgrade to Tika 1.3

2013-03-04 Thread Commit Tag Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4416?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592937#comment-13592937
 ] 

Commit Tag Bot commented on SOLR-4416:
--

[branch_4x commit] Chris M. Hostetter
http://svn.apache.org/viewvc?view=revision&revision=1452629

SOLR-4416: fix SHA files for tika upgrade (merge r1452626)


> Upgrade to Tika 1.3
> ---
>
> Key: SOLR-4416
> URL: https://issues.apache.org/jira/browse/SOLR-4416
> Project: Solr
>  Issue Type: Task
>  Components: contrib - Solr Cell (Tika extraction)
>Affects Versions: 4.1
>Reporter: Erik Hatcher
>Assignee: Mark Miller
>Priority: Critical
> Fix For: 4.2, 5.0
>
> Attachments: SOLR-4416-trunk-1.patch
>
>
> Tika 1.3 was recently released with these improvements: 
> http://www.apache.org/dist/tika/CHANGES-1.3.txt

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-trunk-Linux (64bit/jdk1.6.0_41) - Build # 4573 - Still Failing!

2013-03-04 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-Linux/4573/
Java: 64bit/jdk1.6.0_41 -XX:+UseParallelGC

All tests passed

Build Log:
[...truncated 20437 lines...]
check-licenses:
 [echo] License check under: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/contrib/extraction/lib/fontbox-1.7.1.jar
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/contrib/extraction/lib/jempbox-1.7.1.jar

 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/contrib/extraction/lib/metadata-extractor-2.6.2.jar
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/contrib/extraction/lib/pdfbox-1.7.1.jar
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/contrib/extraction/lib/tika-core-1.3.jar
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/contrib/extraction/lib/tika-parsers-1.3.jar

[...truncated 1 lines...]
BUILD FAILED
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/build.xml:381: The following 
error occurred while executing this line:
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/build.xml:67: The following 
error occurred while executing this line:
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build.xml:234: The 
following error occurred while executing this line:
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/lucene/tools/custom-tasks.xml:43:
 License check failed. Check the logs.

Total time: 43 minutes 51 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Recording test results
Description set: Java: 64bit/jdk1.6.0_41 -XX:+UseParallelGC
Email was triggered for: Failure
Sending email for trigger: Failure



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Commented] (SOLR-4416) Upgrade to Tika 1.3

2013-03-04 Thread Commit Tag Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4416?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592923#comment-13592923
 ] 

Commit Tag Bot commented on SOLR-4416:
--

[trunk commit] Chris M. Hostetter
http://svn.apache.org/viewvc?view=revision&revision=1452626

SOLR-4416: fix SHA files for tika upgrade


> Upgrade to Tika 1.3
> ---
>
> Key: SOLR-4416
> URL: https://issues.apache.org/jira/browse/SOLR-4416
> Project: Solr
>  Issue Type: Task
>  Components: contrib - Solr Cell (Tika extraction)
>Affects Versions: 4.1
>Reporter: Erik Hatcher
>Assignee: Mark Miller
>Priority: Critical
> Fix For: 4.2, 5.0
>
> Attachments: SOLR-4416-trunk-1.patch
>
>
> Tika 1.3 was recently released with these improvements: 
> http://www.apache.org/dist/tika/CHANGES-1.3.txt

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-trunk-MacOSX (64bit/jdk1.7.0) - Build # 289 - Failure!

2013-03-04 Thread Chris Hostetter

FYI: SHA fix in progress... waiting on precommit.



: Date: Tue, 5 Mar 2013 01:00:56 + (UTC)
: From: Policeman Jenkins Server 
: Reply-To: dev@lucene.apache.org
: To: dev@lucene.apache.org, hoss...@apache.org, mikemcc...@apache.org,
: markrmil...@apache.org
: Subject: [JENKINS] Lucene-Solr-trunk-MacOSX (64bit/jdk1.7.0) - Build # 289 -
: Failure!
: 
: Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-MacOSX/289/
: Java: 64bit/jdk1.7.0 -XX:+UseSerialGC
: 
: All tests passed
: 
: Build Log:
: [...truncated 21132 lines...]
: check-licenses:
:  [echo] License check under: 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr
:  [licenses] MISSING sha1 checksum file for: 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/contrib/extraction/lib/fontbox-1.7.1.jar
:  [licenses] MISSING sha1 checksum file for: 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/contrib/extraction/lib/jempbox-1.7.1.jar
: 
:  [licenses] MISSING sha1 checksum file for: 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/contrib/extraction/lib/metadata-extractor-2.6.2.jar
:  [licenses] MISSING sha1 checksum file for: 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/contrib/extraction/lib/pdfbox-1.7.1.jar
:  [licenses] MISSING sha1 checksum file for: 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/contrib/extraction/lib/tika-core-1.3.jar
:  [licenses] MISSING sha1 checksum file for: 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/contrib/extraction/lib/tika-parsers-1.3.jar
: 
: [...truncated 1 lines...]
: BUILD FAILED
: 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/build.xml:381: 
The following error occurred while executing this line:
: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/build.xml:67: 
The following error occurred while executing this line:
: 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build.xml:234:
 The following error occurred while executing this line:
: 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/lucene/tools/custom-tasks.xml:43:
 License check failed. Check the logs.
: 
: Total time: 81 minutes 34 seconds
: Build step 'Invoke Ant' marked build as failure
: Archiving artifacts
: Recording test results
: Description set: Java: 64bit/jdk1.7.0 -XX:+UseSerialGC
: Email was triggered for: Failure
: Sending email for trigger: Failure
: 
: 
: 

-Hoss

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4524) ReturnFields could not support hash '#' in filed name

2013-03-04 Thread Commit Tag Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592897#comment-13592897
 ] 

Commit Tag Bot commented on SOLR-4524:
--

[branch_4x commit] Chris M. Hostetter
http://svn.apache.org/viewvc?view=revision&revision=1452613

SOLR-4524: test proving this works (merge r1452612)


> ReturnFields could not support hash '#' in filed name 
> --
>
> Key: SOLR-4524
> URL: https://issues.apache.org/jira/browse/SOLR-4524
> Project: Solr
>  Issue Type: Bug
>  Components: search
>Affects Versions: 4.1
>Reporter: Benjamin POUSSIN
>Assignee: Hoss Man
>  Labels: field, fl, query, queryparser
>
> If some fields start with '#' and we try to put it in fl param, this field 
> and all next fields are not in response document
> With query:
>   q=*:*&fl=score,#id,myOtherField
> Document response is like:
>SolrDocument{score=3.809621}
> this bug is similar to SOLR-2719 and SOLR-4374

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: Ability to specify 2 different query analyzers for same indexed field in Solr

2013-03-04 Thread Erick Erickson
Tom:

I wonder if you could do something with payloads here. Index all terms with
payloads of 10, but synonyms with 1?

Random thought off the top of my head.

Erick


On Mon, Mar 4, 2013 at 6:25 PM, Tom Burton-West  wrote:

> Hi Jack,
>
> Sorry the example is not clear.  Below is the normal way to accomplish
> what I am trying to do using a copyField and two separate fieldTypes with
> the index analyzer the same but the query time analyzer different.
>
> So the query would be something like q=plain:foobar^10 OR syn:foobar^1  to
> get synonyms but scored much lower than an exact match.
>
> The problem with this is since the analysis chain used for indexing is the
> same in both cases, I would rather not have to actually index the exact
> same content in the exact same way twice.
>
> Does that make it any clearer or do I need a more compelling use case?
>
> Tom
>
> 
> 
>
>   
> 
> 
> 
>
>   
> 
>
> 
> 
>
>   
> 
> 
> 
>
> ignoreCase="true" expand="true"/>
>   
> 
> 
>
> On Mon, Mar 4, 2013 at 4:43 PM, Jack Krupansky wrote:
>
>>   Please clarify, and try providing a couple more use cases. I mean, the
>> case you provided suggests that the contents of the index will be different
>> between the two fields, while you told us that you wanted to share the same
>> indexed field. In other words, it sounds like you will have two copies of
>> similar data anyway.
>>
>> Maybe you simply want one copy of the stored value for the field and then
>> have one or more copyfields that index the same source data differently,
>> but don’t re-store the copied source data.
>>
>> -- Jack Krupansky
>>
>>  *From:* Tom Burton-West 
>> *Sent:* Monday, March 04, 2013 3:57 PM
>> *To:* dev@lucene.apache.org
>> *Subject:* Ability to specify 2 different query analyzers for same
>> indexed field in Solr
>>
>> Hello,
>>
>> We would like to be able to specify two different fields that both use
>> the same indexed field but use different analyzers.   An example use-case
>> for this might be doing query-time synonym expansion with the synonyms
>> weighted lower than an exact match.
>>
>> q=exact_field^10 OR synonyms^1
>>
>> The normal way to do this in Solr, which is just to set up separate
>> analyzer chains and use a copyfield, will not work for us because the field
>> in question is huge.  It is about 7 TB of OCR.
>>
>> Is there a way to do this currently in Solr?   If not ,
>>
>> 1) should I open a JIRA issue?
>> 2) can someone point me towards the part of the code I might need to
>> modify?
>>
>> Tom
>>
>>  Tom Burton-West
>> Information Retrieval Programmer
>> Digital Library Production Service
>> University of Michigan Library
>> http://www.hathitrust.org/blogs/large-scale-search
>>
>>
>>
>
>


[jira] [Comment Edited] (SOLR-4416) Upgrade to Tika 1.3

2013-03-04 Thread Hoss Man (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4416?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592892#comment-13592892
 ] 

Hoss Man edited comment on SOLR-4416 at 3/5/13 1:17 AM:


This is causing build failures because of SHA & license checking on -transient- 
jars

EDIT: the problem isn't actually transitive dependencies, the jars with SHA 
problems are explicitly listed, but their SHA files weren't updated.

  was (Author: hossman):
This is causing build failures because of SHA & license checking on 
transient jars
  
> Upgrade to Tika 1.3
> ---
>
> Key: SOLR-4416
> URL: https://issues.apache.org/jira/browse/SOLR-4416
> Project: Solr
>  Issue Type: Task
>  Components: contrib - Solr Cell (Tika extraction)
>Affects Versions: 4.1
>Reporter: Erik Hatcher
>Assignee: Mark Miller
>Priority: Critical
> Fix For: 4.2, 5.0
>
> Attachments: SOLR-4416-trunk-1.patch
>
>
> Tika 1.3 was recently released with these improvements: 
> http://www.apache.org/dist/tika/CHANGES-1.3.txt

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Reopened] (SOLR-4416) Upgrade to Tika 1.3

2013-03-04 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4416?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man reopened SOLR-4416:



This is causing build failures because of SHA & license checking on transient 
jars

> Upgrade to Tika 1.3
> ---
>
> Key: SOLR-4416
> URL: https://issues.apache.org/jira/browse/SOLR-4416
> Project: Solr
>  Issue Type: Task
>  Components: contrib - Solr Cell (Tika extraction)
>Affects Versions: 4.1
>Reporter: Erik Hatcher
>Assignee: Mark Miller
>Priority: Critical
> Fix For: 4.2, 5.0
>
> Attachments: SOLR-4416-trunk-1.patch
>
>
> Tika 1.3 was recently released with these improvements: 
> http://www.apache.org/dist/tika/CHANGES-1.3.txt

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-4524) ReturnFields could not support hash '#' in filed name

2013-03-04 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4524?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man resolved SOLR-4524.


Resolution: Cannot Reproduce

Tests added...

Committed revision 1452612.
Committed revision 1452613.


> ReturnFields could not support hash '#' in filed name 
> --
>
> Key: SOLR-4524
> URL: https://issues.apache.org/jira/browse/SOLR-4524
> Project: Solr
>  Issue Type: Bug
>  Components: search
>Affects Versions: 4.1
>Reporter: Benjamin POUSSIN
>Assignee: Hoss Man
>  Labels: field, fl, query, queryparser
>
> If some fields start with '#' and we try to put it in fl param, this field 
> and all next fields are not in response document
> With query:
>   q=*:*&fl=score,#id,myOtherField
> Document response is like:
>SolrDocument{score=3.809621}
> this bug is similar to SOLR-2719 and SOLR-4374

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4524) ReturnFields could not support hash '#' in filed name

2013-03-04 Thread Commit Tag Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4524?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592885#comment-13592885
 ] 

Commit Tag Bot commented on SOLR-4524:
--

[trunk commit] Chris M. Hostetter
http://svn.apache.org/viewvc?view=revision&revision=1452612

SOLR-4524: test proving this works


> ReturnFields could not support hash '#' in filed name 
> --
>
> Key: SOLR-4524
> URL: https://issues.apache.org/jira/browse/SOLR-4524
> Project: Solr
>  Issue Type: Bug
>  Components: search
>Affects Versions: 4.1
>Reporter: Benjamin POUSSIN
>Assignee: Hoss Man
>  Labels: field, fl, query, queryparser
>
> If some fields start with '#' and we try to put it in fl param, this field 
> and all next fields are not in response document
> With query:
>   q=*:*&fl=score,#id,myOtherField
> Document response is like:
>SolrDocument{score=3.809621}
> this bug is similar to SOLR-2719 and SOLR-4374

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-4.x-Linux (32bit/jdk1.7.0_15) - Build # 4542 - Failure!

2013-03-04 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/4542/
Java: 32bit/jdk1.7.0_15 -client -XX:+UseSerialGC

All tests passed

Build Log:
[...truncated 21252 lines...]
check-licenses:
 [echo] License check under: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/contrib/extraction/lib/fontbox-1.7.1.jar
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/contrib/extraction/lib/jempbox-1.7.1.jar

 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/contrib/extraction/lib/metadata-extractor-2.6.2.jar
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/contrib/extraction/lib/pdfbox-1.7.1.jar
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/contrib/extraction/lib/tika-core-1.3.jar
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/contrib/extraction/lib/tika-parsers-1.3.jar

[...truncated 1 lines...]
BUILD FAILED
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/build.xml:381: The following 
error occurred while executing this line:
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/build.xml:67: The following 
error occurred while executing this line:
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/build.xml:234: The 
following error occurred while executing this line:
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/tools/custom-tasks.xml:43:
 License check failed. Check the logs.

Total time: 61 minutes 7 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Recording test results
Description set: Java: 32bit/jdk1.7.0_15 -client -XX:+UseSerialGC
Email was triggered for: Failure
Sending email for trigger: Failure



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[JENKINS] Lucene-Solr-trunk-MacOSX (64bit/jdk1.7.0) - Build # 289 - Failure!

2013-03-04 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-MacOSX/289/
Java: 64bit/jdk1.7.0 -XX:+UseSerialGC

All tests passed

Build Log:
[...truncated 21132 lines...]
check-licenses:
 [echo] License check under: 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr
 [licenses] MISSING sha1 checksum file for: 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/contrib/extraction/lib/fontbox-1.7.1.jar
 [licenses] MISSING sha1 checksum file for: 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/contrib/extraction/lib/jempbox-1.7.1.jar

 [licenses] MISSING sha1 checksum file for: 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/contrib/extraction/lib/metadata-extractor-2.6.2.jar
 [licenses] MISSING sha1 checksum file for: 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/contrib/extraction/lib/pdfbox-1.7.1.jar
 [licenses] MISSING sha1 checksum file for: 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/contrib/extraction/lib/tika-core-1.3.jar
 [licenses] MISSING sha1 checksum file for: 
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/contrib/extraction/lib/tika-parsers-1.3.jar

[...truncated 1 lines...]
BUILD FAILED
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/build.xml:381: 
The following error occurred while executing this line:
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/build.xml:67: 
The following error occurred while executing this line:
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build.xml:234:
 The following error occurred while executing this line:
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/lucene/tools/custom-tasks.xml:43:
 License check failed. Check the logs.

Total time: 81 minutes 34 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Recording test results
Description set: Java: 64bit/jdk1.7.0 -XX:+UseSerialGC
Email was triggered for: Failure
Sending email for trigger: Failure



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Comment Edited] (LUCENE-4694) Add back IndexReader.fields() -> Multi*, or discourage term vectors in some better way

2013-03-04 Thread Uwe Schindler (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4694?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592819#comment-13592819
 ] 

Uwe Schindler edited comment on LUCENE-4694 at 3/5/13 12:03 AM:


Mike: If you really, really want to move it up:
- Make fields() abstract in IndexReader, only add the final methods that 
delegate to the abstract one
- leave fields() abstract in AtomicReader and CompositeReader, too
- add the MultiFields implementation into BaseCompositeReader -> and only there 
(maybe final)
- the javadocs with warning are only in CompositeReader

The current patch is as worse as IndexReader before the split into 
composite/atomic! IndexReader is an abstract class without concrete 
implementations. The final methods only delegate to abstract methods, so they 
are no "implementations".

But I don't like the whole thing in general. Just because few users are doing 
those thing, we should never ever expose postings on composite readers. In my 
opinion, the term vectors should simply go away?

Stored fields with the binary search are only there to make "result display" 
possible. Maybe we should remove stored fields from IndexReader base class, 
too. Instead stored field are only accessible through IndexSearcher (which 
delegates using its reader contexts)?

  was (Author: thetaphi):
Mike: If you really, really want to move it up:
- Make fields() abstract in IndexReader, only add the final methods that 
delegate to the abstract one
- leave fields() abstract in AtomicReader and CompositeReader, too
- add the MultiFields implementation into BaseCompositeReader -> and only there 
(maybe final)

The current patch is as worse as IndexReader before the split into 
composite/atomic! IndexReader is an abstract class without concrete 
implementations. The final methods only delegate to abstract methods, so they 
are no "implementations".
  
> Add back IndexReader.fields() -> Multi*, or discourage term vectors in some 
> better way
> --
>
> Key: LUCENE-4694
> URL: https://issues.apache.org/jira/browse/LUCENE-4694
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Robert Muir
> Attachments: LUCENE-4694.patch
>
>
> Users can easily get term vectors from any indexreader, but not postings 
> lists. this encourages them to do really slow things: like pulling term 
> vectors for every single document.
> this is really really so much worse than going through multifields or 
> whatever. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4694) Add back IndexReader.fields() -> Multi*, or discourage term vectors in some better way

2013-03-04 Thread Uwe Schindler (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4694?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592819#comment-13592819
 ] 

Uwe Schindler commented on LUCENE-4694:
---

Mike: If you really, really want to move it up:
- Make fields() abstract in IndexReader, only add the final methods that 
delegate to the abstract one
- leave fields() abstract in AtomicReader and CompositeReader, too
- add the MultiFields implementation into BaseCompositeReader -> and only there 
(maybe final)

The current patch is as worse as IndexReader before the split into 
composite/atomic! IndexReader is an abstract class without concrete 
implementations. The final methods only delegate to abstract methods, so they 
are no "implementations".

> Add back IndexReader.fields() -> Multi*, or discourage term vectors in some 
> better way
> --
>
> Key: LUCENE-4694
> URL: https://issues.apache.org/jira/browse/LUCENE-4694
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Robert Muir
> Attachments: LUCENE-4694.patch
>
>
> Users can easily get term vectors from any indexreader, but not postings 
> lists. this encourages them to do really slow things: like pulling term 
> vectors for every single document.
> this is really really so much worse than going through multifields or 
> whatever. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS-MAVEN] Lucene-Solr-Maven-trunk #790: POMs out of sync

2013-03-04 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-Maven-trunk/790/

1 tests failed.
REGRESSION:  org.apache.solr.cloud.SyncSliceTest.testDistribSearch

Error Message:
Test Setup Failure: shard1 should have just been set up to be inconsistent - 
but it's still consistent. Leader:http://127.0.0.1:35140/zv/collection1skip 
list:[CloudJettyRunner [url=http://127.0.0.1:28486/zv/collection1], 
CloudJettyRunner [url=http://127.0.0.1:28486/zv/collection1]]

Stack Trace:
java.lang.AssertionError: Test Setup Failure: shard1 should have just been set 
up to be inconsistent - but it's still consistent. 
Leader:http://127.0.0.1:35140/zv/collection1skip list:[CloudJettyRunner 
[url=http://127.0.0.1:28486/zv/collection1], CloudJettyRunner 
[url=http://127.0.0.1:28486/zv/collection1]]
at 
__randomizedtesting.SeedInfo.seed([4510B6D6F811A0C3:C4F638CE8F4EC0FF]:0)
at org.junit.Assert.fail(Assert.java:93)
at org.junit.Assert.assertTrue(Assert.java:43)
at org.junit.Assert.assertNotNull(Assert.java:526)
at org.apache.solr.cloud.SyncSliceTest.doTest(SyncSliceTest.java:212)




Build Log:
[...truncated 22634 lines...]



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Commented] (LUCENE-4694) Add back IndexReader.fields() -> Multi*, or discourage term vectors in some better way

2013-03-04 Thread Uwe Schindler (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4694?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592812#comment-13592812
 ] 

Uwe Schindler commented on LUCENE-4694:
---

Oh my god :(

> Add back IndexReader.fields() -> Multi*, or discourage term vectors in some 
> better way
> --
>
> Key: LUCENE-4694
> URL: https://issues.apache.org/jira/browse/LUCENE-4694
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Robert Muir
> Attachments: LUCENE-4694.patch
>
>
> Users can easily get term vectors from any indexreader, but not postings 
> lists. this encourages them to do really slow things: like pulling term 
> vectors for every single document.
> this is really really so much worse than going through multifields or 
> whatever. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (LUCENE-4694) Add back IndexReader.fields() -> Multi*, or discourage term vectors in some better way

2013-03-04 Thread Michael McCandless (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4694?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Michael McCandless updated LUCENE-4694:
---

Attachment: LUCENE-4694.patch

Patch, moving the postings APIs up from AR to IR, and calling 
MultiFields.getFields by default.

> Add back IndexReader.fields() -> Multi*, or discourage term vectors in some 
> better way
> --
>
> Key: LUCENE-4694
> URL: https://issues.apache.org/jira/browse/LUCENE-4694
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Robert Muir
> Attachments: LUCENE-4694.patch
>
>
> Users can easily get term vectors from any indexreader, but not postings 
> lists. this encourages them to do really slow things: like pulling term 
> vectors for every single document.
> this is really really so much worse than going through multifields or 
> whatever. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: Ability to specify 2 different query analyzers for same indexed field in Solr

2013-03-04 Thread Tom Burton-West
Hi Jack,

Sorry the example is not clear.  Below is the normal way to accomplish what
I am trying to do using a copyField and two separate fieldTypes with the
index analyzer the same but the query time analyzer different.

So the query would be something like q=plain:foobar^10 OR syn:foobar^1  to
get synonyms but scored much lower than an exact match.

The problem with this is since the analysis chain used for indexing is the
same in both cases, I would rather not have to actually index the exact
same content in the exact same way twice.

Does that make it any clearer or do I need a more compelling use case?

Tom



   
  



   
  




   
  



   
   
  



On Mon, Mar 4, 2013 at 4:43 PM, Jack Krupansky wrote:

>   Please clarify, and try providing a couple more use cases. I mean, the
> case you provided suggests that the contents of the index will be different
> between the two fields, while you told us that you wanted to share the same
> indexed field. In other words, it sounds like you will have two copies of
> similar data anyway.
>
> Maybe you simply want one copy of the stored value for the field and then
> have one or more copyfields that index the same source data differently,
> but don’t re-store the copied source data.
>
> -- Jack Krupansky
>
>  *From:* Tom Burton-West 
> *Sent:* Monday, March 04, 2013 3:57 PM
> *To:* dev@lucene.apache.org
> *Subject:* Ability to specify 2 different query analyzers for same
> indexed field in Solr
>
> Hello,
>
> We would like to be able to specify two different fields that both use the
> same indexed field but use different analyzers.   An example use-case for
> this might be doing query-time synonym expansion with the synonyms weighted
> lower than an exact match.
>
> q=exact_field^10 OR synonyms^1
>
> The normal way to do this in Solr, which is just to set up separate
> analyzer chains and use a copyfield, will not work for us because the field
> in question is huge.  It is about 7 TB of OCR.
>
> Is there a way to do this currently in Solr?   If not ,
>
> 1) should I open a JIRA issue?
> 2) can someone point me towards the part of the code I might need to
> modify?
>
> Tom
>
>  Tom Burton-West
> Information Retrieval Programmer
> Digital Library Production Service
> University of Michigan Library
> http://www.hathitrust.org/blogs/large-scale-search
>
>
>


[JENKINS] Lucene-Solr-trunk-Linux (32bit/jdk1.6.0_41) - Build # 4571 - Still Failing!

2013-03-04 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-Linux/4571/
Java: 32bit/jdk1.6.0_41 -client -XX:+UseParallelGC

All tests passed

Build Log:
[...truncated 20411 lines...]
check-licenses:
 [echo] License check under: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/contrib/extraction/lib/fontbox-1.7.1.jar
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/contrib/extraction/lib/jempbox-1.7.1.jar

 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/contrib/extraction/lib/metadata-extractor-2.6.2.jar
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/contrib/extraction/lib/pdfbox-1.7.1.jar
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/contrib/extraction/lib/tika-core-1.3.jar
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/contrib/extraction/lib/tika-parsers-1.3.jar

[...truncated 1 lines...]
BUILD FAILED
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/build.xml:381: The following 
error occurred while executing this line:
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/build.xml:67: The following 
error occurred while executing this line:
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build.xml:234: The 
following error occurred while executing this line:
/mnt/ssd/jenkins/workspace/Lucene-Solr-trunk-Linux/lucene/tools/custom-tasks.xml:43:
 License check failed. Check the logs.

Total time: 50 minutes 19 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Recording test results
Description set: Java: 32bit/jdk1.6.0_41 -client -XX:+UseParallelGC
Email was triggered for: Failure
Sending email for trigger: Failure



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Closed] (SOLR-600) XML parser stops working under heavy load

2013-03-04 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/SOLR-600?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jan Høydahl closed SOLR-600.


Resolution: Cannot Reproduce

Closing as this seems not to be an issue with current versions. Please re-open 
if you discover this again.

> XML parser stops working under heavy load
> -
>
> Key: SOLR-600
> URL: https://issues.apache.org/jira/browse/SOLR-600
> Project: Solr
>  Issue Type: Bug
>  Components: update
>Affects Versions: 1.3
> Environment: Linux 2.6.19.7-ss0 #4 SMP Wed Mar 12 02:56:42 GMT 2008 
> x86_64 Intel(R) Xeon(R) CPU X5450 @ 3.00GHz GenuineIntel GNU/Linux
> Tomcat 6.0.16
> SOLR nightly 16 Jun 2008, and versions prior
> JRE 1.6.0
>Reporter: John Smith
>
> Under heavy load, the following is spat out for every update:
> org.apache.solr.common.SolrException log
> SEVERE: java.lang.NullPointerException
> at java.util.AbstractList$SimpleListIterator.hasNext(Unknown Source)
> at 
> org.apache.solr.update.DocumentBuilder.toDocument(DocumentBuilder.java:225)
> at 
> org.apache.solr.update.processor.RunUpdateProcessor.processAdd(RunUpdateProcessorFactory.java:66)
> at 
> org.apache.solr.handler.XmlUpdateRequestHandler.processUpdate(XmlUpdateRequestHandler.java:196)
> at 
> org.apache.solr.handler.XmlUpdateRequestHandler.handleRequestBody(XmlUpdateRequestHandler.java:123)
> at 
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:125)
> at org.apache.solr.core.SolrCore.execute(SolrCore.java:965)
> at 
> org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338)
> at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:272)
> at 
> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:235)
> at 
> org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
> at 
> org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:233)
> at 
> org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:175)
> at 
> org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:128)
> at 
> org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102)
> at 
> org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:109)
> at 
> org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:286)
> at 
> org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:844)
> at 
> org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http11Protocol.java:583)
> at 
> org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:447)
> at java.lang.Thread.run(Thread.java:735)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3459) Change ChainedFilter to use FixedBitSet

2013-03-04 Thread Uwe Schindler (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3459?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592746#comment-13592746
 ] 

Uwe Schindler commented on LUCENE-3459:
---

If you look into the details of usage, you will see recognize that OpenBitSet 
is only usd by Solr. Lucene does not use it anymore except at these few places. 
ChainedFilter needs some additional bit operations, FixedBitSet currently does 
not support.

Performance-wise FixedBitSet is better for filters because it has less checks 
and does not support resize (the bit set size is fixed). 

> Change ChainedFilter to use FixedBitSet
> ---
>
> Key: LUCENE-3459
> URL: https://issues.apache.org/jira/browse/LUCENE-3459
> Project: Lucene - Core
>  Issue Type: Task
>  Components: modules/other
>Affects Versions: 3.4, 4.0-ALPHA
>Reporter: Uwe Schindler
>Assignee: Uwe Schindler
> Fix For: 4.2
>
>
> ChainedFilter also uses OpenBitSet(DISI) at the moment. It should also be 
> changed to use FixedBitSet. There are two issues:
> - It exposes sometimes OpenBitSetDISI to it's public API - we should remove 
> those methods like in BooleanFilter and break backwards
> - It allows a XOR operation. This is not yet supported by FixedBitSet, but 
> it's easy to add (like for BooleanFilter). On the other hand, this XOR 
> operation is bogus, as it may mark documents in the BitSet that are deleted, 
> breaking new features like applying Filters down-low (LUCENE-1536). We should 
> remove the XOR operation maybe or force it to use IR.validDocs() (trunk) or 
> IR.isDeleted()

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3459) Change ChainedFilter to use FixedBitSet

2013-03-04 Thread David Smiley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3459?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592740#comment-13592740
 ] 

David Smiley commented on LUCENE-3459:
--

Ha!  I was just about to report the reverse problem -- make TermsFilter return 
OpenBitset for optimal compatibility with ChainedFilter.  I'm using 
TermsFilters with ChainedFilter and the mis-match of OpenBitSet and FixedBitSet 
is definitely not optimal since I want it to intersect those bitsets together 
quickly.

I don't understand why there are both.  My IDE says FixedBitSet has 206 usages 
compared to OpenBitSet's 230.  That's closer than I thought.  I recall 
[~thetaphi‍] saying he would like OBS deprecated, but I pointed out that it is 
OBS that is fully public and FixedBitSet that is marked as @lucene.internal.

> Change ChainedFilter to use FixedBitSet
> ---
>
> Key: LUCENE-3459
> URL: https://issues.apache.org/jira/browse/LUCENE-3459
> Project: Lucene - Core
>  Issue Type: Task
>  Components: modules/other
>Affects Versions: 3.4, 4.0-ALPHA
>Reporter: Uwe Schindler
>Assignee: Uwe Schindler
> Fix For: 4.2
>
>
> ChainedFilter also uses OpenBitSet(DISI) at the moment. It should also be 
> changed to use FixedBitSet. There are two issues:
> - It exposes sometimes OpenBitSetDISI to it's public API - we should remove 
> those methods like in BooleanFilter and break backwards
> - It allows a XOR operation. This is not yet supported by FixedBitSet, but 
> it's easy to add (like for BooleanFilter). On the other hand, this XOR 
> operation is bogus, as it may mark documents in the BitSet that are deleted, 
> breaking new features like applying Filters down-low (LUCENE-1536). We should 
> remove the XOR operation maybe or force it to use IR.validDocs() (trunk) or 
> IR.isDeleted()

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-4.x-Linux (64bit/jdk1.7.0_15) - Build # 4540 - Failure!

2013-03-04 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/4540/
Java: 64bit/jdk1.7.0_15 -XX:+UseConcMarkSweepGC

All tests passed

Build Log:
[...truncated 21446 lines...]
check-licenses:
 [echo] License check under: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/contrib/extraction/lib/fontbox-1.7.1.jar
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/contrib/extraction/lib/jempbox-1.7.1.jar

 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/contrib/extraction/lib/metadata-extractor-2.6.2.jar
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/contrib/extraction/lib/pdfbox-1.7.1.jar
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/contrib/extraction/lib/tika-core-1.3.jar
 [licenses] MISSING sha1 checksum file for: 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/contrib/extraction/lib/tika-parsers-1.3.jar

[...truncated 1 lines...]
BUILD FAILED
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/build.xml:381: The following 
error occurred while executing this line:
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/build.xml:67: The following 
error occurred while executing this line:
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/solr/build.xml:234: The 
following error occurred while executing this line:
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/tools/custom-tasks.xml:43:
 License check failed. Check the logs.

Total time: 48 minutes 14 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Recording test results
Description set: Java: 64bit/jdk1.7.0_15 -XX:+UseConcMarkSweepGC
Email was triggered for: Failure
Sending email for trigger: Failure



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Commented] (LUCENE-4810) Positions are incremented for each ngram in EdgeNGramTokenFilter

2013-03-04 Thread David Smiley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4810?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592711#comment-13592711
 ] 

David Smiley commented on LUCENE-4810:
--

+1

> Positions are incremented for each ngram in EdgeNGramTokenFilter
> 
>
> Key: LUCENE-4810
> URL: https://issues.apache.org/jira/browse/LUCENE-4810
> Project: Lucene - Core
>  Issue Type: Bug
>  Components: modules/analysis
>Reporter: Walter Underwood
> Attachments: LUCENE-4810.diff
>
>
> Edge ngrams should be like synonyms, with all the ngrams generated from a 
> token having the same position as that original token. The current code 
> increments position.
> For the text "molecular biology", the query "mol bio" should match as a 
> phrase in neighboring positions. It does not.
> You can see this in the Analysis page in the admin UI.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4526) Admin UI dependence on UnixOperationSystemMXBean

2013-03-04 Thread Uwe Schindler (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4526?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592685#comment-13592685
 ] 

Uwe Schindler commented on SOLR-4526:
-

The general problem here is the admin UI, as mentioned by Felix. The output of 
the MX data is very JVM-specific, so the UI should not rely on key/value pairs 
are always available. If you e.g. start Solr with IBM J9, the whole Andmin UI 
shows nothing at all for those statistics, same for JRockit. If you look into 
browser JS logs you see errors about missing keys.

The RequestHandler on the Solr side is fine and produces no log entries. 
Because of that you will see no errors in the log file, as the UI cannot log 
anything to the Solr log.

The reason why the Ubuntu-Jetty does not display those information might be 
because the Webapp context is restricted and does not provide the OS-MXBean. 
The Jetty included with Solr has no security settings at all and the webapp is 
running in the root context.

> Admin UI dependence on UnixOperationSystemMXBean
> 
>
> Key: SOLR-4526
> URL: https://issues.apache.org/jira/browse/SOLR-4526
> Project: Solr
>  Issue Type: Bug
>  Components: web gui
>Affects Versions: 4.1
> Environment: Oracle Corporation OpenJDK 64-Bit Server VM (1.7.0_15 
> 23.7-b01)
>Reporter: Shawn Heisey
>Assignee: Stefan Matheis (steffkes)
> Attachments: built-in-jetty8.png, ubuntu-jetty6.png
>
>
> A user on IRC was having trouble getting file descriptor counts and JVM 
> memory usage in the admin UI, but it worked perfectly fine on another system. 
>  The problem system uses OpenJDK, the other one uses the Apple JDK.  The user 
> had tracked it down to an exception while trying to get open file descriptor 
> info.
> Looking in the SystemInfoHandler.java file, I see a comment reference to 
> com.sun.management.UnixOperatingSystemMXBean at the point where it is getting 
> file descriptor info.  A little extra searching turned up ZOOKEEPER-1579 
> which refers to HBASE-6945, the same problem with OpenJDK.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-trunk-Windows (64bit/jdk1.6.0_41) - Build # 2618 - Still Failing!

2013-03-04 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-Windows/2618/
Java: 64bit/jdk1.6.0_41 -XX:+UseConcMarkSweepGC

All tests passed

Build Log:
[...truncated 28646 lines...]
BUILD FAILED
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\build.xml:381: The 
following error occurred while executing this line:
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\build.xml:60: The 
following error occurred while executing this line:
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\solr\build.xml:591: 
The following error occurred while executing this line:
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\lucene\common-build.xml:1591:
 The following error occurred while executing this line:
C:\Users\JenkinsSlave\workspace\Lucene-Solr-trunk-Windows\lucene\common-build.xml:1625:
 Compile failed; see the compiler error output for details.

Total time: 72 minutes 31 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Recording test results
Description set: Java: 64bit/jdk1.6.0_41 -XX:+UseConcMarkSweepGC
Email was triggered for: Failure
Sending email for trigger: Failure



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

Re: Ability to specify 2 different query analyzers for same indexed field in Solr

2013-03-04 Thread Jack Krupansky
Please clarify, and try providing a couple more use cases. I mean, the case you 
provided suggests that the contents of the index will be different between the 
two fields, while you told us that you wanted to share the same indexed field. 
In other words, it sounds like you will have two copies of similar data anyway.

Maybe you simply want one copy of the stored value for the field and then have 
one or more copyfields that index the same source data differently, but don’t 
re-store the copied source data.

-- Jack Krupansky

From: Tom Burton-West 
Sent: Monday, March 04, 2013 3:57 PM
To: dev@lucene.apache.org 
Subject: Ability to specify 2 different query analyzers for same indexed field 
in Solr

Hello, 

We would like to be able to specify two different fields that both use the same 
indexed field but use different analyzers.   An example use-case for this might 
be doing query-time synonym expansion with the synonyms weighted lower than an 
exact match.   

q=exact_field^10 OR synonyms^1

The normal way to do this in Solr, which is just to set up separate analyzer 
chains and use a copyfield, will not work for us because the field in question 
is huge.  It is about 7 TB of OCR.

Is there a way to do this currently in Solr?   If not ,

1) should I open a JIRA issue?
2) can someone point me towards the part of the code I might need to modify?

Tom 

Tom Burton-West
Information Retrieval Programmer
Digital Library Production Service
University of Michigan Library
http://www.hathitrust.org/blogs/large-scale-search



[jira] [Comment Edited] (SOLR-3798) copyField logic in LukeRequestHandler is primitive, doesn't work well with dynamicFields

2013-03-04 Thread Steve Rowe (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3798?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592665#comment-13592665
 ] 

Steve Rowe edited comment on SOLR-3798 at 3/4/13 9:43 PM:
--

In the latest patch on SOLR-4503, I've included a dynamic copy field 
refactoring in IndexSchema.java that fixes cases 7, 11, 14, and 15 from the 
above table - with that patch, the Solr schema parse succeeds for those cases.

But LukeRequestHandler's response can only carry those cases where at least one 
of the source or the dest is a declared field or dynamic field name, so its 
current implementation won't handle cases 11, 12, 15, or 16.  I think 
LukeRequestHandler should split out copyField info, something like the way I 
did it on SOLR-4503, so that subset patterns can be reported.

  was (Author: steve_rowe):
In the latest patch on SOLR-4503, I've included a dynamic copy field 
refactoring in IndexSchema.java that fixes cases 7, 11, 14, and 15 from the 
above table - with that patch, the Solr schema parse succeeds for those cases.

But LukeRequestHandler's response can only carry those cases where at least one 
of the source or the dest is a declared field or dynamic field name, so it's 
current implementation won't handle cases 11, 12, 15, or 16.  I think 
LukeRequestHandler should split out copyField info, something like the way I 
did it on SOLR-4503, so that subset patterns can be reported.
  
> copyField logic in LukeRequestHandler is primitive, doesn't work well with 
> dynamicFields
> 
>
> Key: SOLR-3798
> URL: https://issues.apache.org/jira/browse/SOLR-3798
> Project: Solr
>  Issue Type: Bug
>Reporter: Hoss Man
> Attachments: SOLR-3798.patch
>
>
> looking into SOLR-3795 i realized there is a much bigger problem with how 
> LukeRequestHandler tries to get copyfield info for fields and dynamicFields 
> the same way, and it just doesn't work.
> see the patch in SOLR-3795 for a commented out example of a test that still 
> fails (ie: trying to get the "copySource" info for a dynamicField)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-3798) copyField logic in LukeRequestHandler is primitive, doesn't work well with dynamicFields

2013-03-04 Thread Steve Rowe (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-3798?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592665#comment-13592665
 ] 

Steve Rowe commented on SOLR-3798:
--

In the latest patch on SOLR-4503, I've included a dynamic copy field 
refactoring in IndexSchema.java that fixes cases 7, 11, 14, and 15 from the 
above table - with that patch, the Solr schema parse succeeds for those cases.

But LukeRequestHandler's response can only carry those cases where at least one 
of the source or the dest is a declared field or dynamic field name, so it's 
current implementation won't handle cases 11, 12, 15, or 16.  I think 
LukeRequestHandler should split out copyField info, something like the way I 
did it on SOLR-4503, so that subset patterns can be reported.

> copyField logic in LukeRequestHandler is primitive, doesn't work well with 
> dynamicFields
> 
>
> Key: SOLR-3798
> URL: https://issues.apache.org/jira/browse/SOLR-3798
> Project: Solr
>  Issue Type: Bug
>Reporter: Hoss Man
> Attachments: SOLR-3798.patch
>
>
> looking into SOLR-3795 i realized there is a much bigger problem with how 
> LukeRequestHandler tries to get copyfield info for fields and dynamicFields 
> the same way, and it just doesn't work.
> see the patch in SOLR-3795 for a commented out example of a test that still 
> fails (ie: trying to get the "copySource" info for a dynamicField)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4526) Admin UI dependence on UnixOperationSystemMXBean

2013-03-04 Thread Felix Buenemann (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4526?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592661#comment-13592661
 ] 

Felix Buenemann commented on SOLR-4526:
---

Hoss: Solr is not logging any errors. I've tried with logging for solr and 
children set to INFO.

> Admin UI dependence on UnixOperationSystemMXBean
> 
>
> Key: SOLR-4526
> URL: https://issues.apache.org/jira/browse/SOLR-4526
> Project: Solr
>  Issue Type: Bug
>  Components: web gui
>Affects Versions: 4.1
> Environment: Oracle Corporation OpenJDK 64-Bit Server VM (1.7.0_15 
> 23.7-b01)
>Reporter: Shawn Heisey
>Assignee: Stefan Matheis (steffkes)
> Attachments: built-in-jetty8.png, ubuntu-jetty6.png
>
>
> A user on IRC was having trouble getting file descriptor counts and JVM 
> memory usage in the admin UI, but it worked perfectly fine on another system. 
>  The problem system uses OpenJDK, the other one uses the Apple JDK.  The user 
> had tracked it down to an exception while trying to get open file descriptor 
> info.
> Looking in the SystemInfoHandler.java file, I see a comment reference to 
> com.sun.management.UnixOperatingSystemMXBean at the point where it is getting 
> file descriptor info.  A little extra searching turned up ZOOKEEPER-1579 
> which refers to HBASE-6945, the same problem with OpenJDK.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4138) currency field doesn't work with functions (ie: isn't compatible with frange query)

2013-03-04 Thread Commit Tag Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4138?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592650#comment-13592650
 ] 

Commit Tag Bot commented on SOLR-4138:
--

[trunk commit] Chris M. Hostetter
http://svn.apache.org/viewvc?view=revision&revision=1452508

SOLR-4138: doc typos


> currency field doesn't work with functions (ie: isn't compatible with frange 
> query)
> ---
>
> Key: SOLR-4138
> URL: https://issues.apache.org/jira/browse/SOLR-4138
> Project: Solr
>  Issue Type: Bug
>  Components: search
>Affects Versions: 4.0
>Reporter: Grzegorz Sobczyk
>Assignee: Hoss Man
> Fix For: 4.2, 5.0
>
> Attachments: SOLR-4135-test.patch, SOLR-4138.patch, SOLR-4138.patch
>
>
> In general, using CurrencyField with FunctionQueries doesn't work
> In particular, as originally reported...
> Filtering using {!frange} syntax isn't work properly. (rather at all)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4138) currency field doesn't work with functions (ie: isn't compatible with frange query)

2013-03-04 Thread Commit Tag Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4138?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592649#comment-13592649
 ] 

Commit Tag Bot commented on SOLR-4138:
--

[branch_4x commit] Chris M. Hostetter
http://svn.apache.org/viewvc?view=revision&revision=1452527

SOLR-4138: CurrencyField fields can now be used in a ValueSources. There is 
also a new currency(field,[CODE]) function (merge r1452483 and r1452508)


> currency field doesn't work with functions (ie: isn't compatible with frange 
> query)
> ---
>
> Key: SOLR-4138
> URL: https://issues.apache.org/jira/browse/SOLR-4138
> Project: Solr
>  Issue Type: Bug
>  Components: search
>Affects Versions: 4.0
>Reporter: Grzegorz Sobczyk
>Assignee: Hoss Man
> Fix For: 4.2, 5.0
>
> Attachments: SOLR-4135-test.patch, SOLR-4138.patch, SOLR-4138.patch
>
>
> In general, using CurrencyField with FunctionQueries doesn't work
> In particular, as originally reported...
> Filtering using {!frange} syntax isn't work properly. (rather at all)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4526) Admin UI dependence on UnixOperationSystemMXBean

2013-03-04 Thread Felix Buenemann (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4526?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592648#comment-13592648
 ] 

Felix Buenemann commented on SOLR-4526:
---

Hoss: I'll try to get my VM back into the prober state and check it.

Regarding what the JS could do better: It would be a good idea to check if the 
accessed array keys are undefined, to avoid an exception. This would make the 
JS code more verbose though.
Had the code checked for undefined values, it would have shown the JVM usage 
just fine.

> Admin UI dependence on UnixOperationSystemMXBean
> 
>
> Key: SOLR-4526
> URL: https://issues.apache.org/jira/browse/SOLR-4526
> Project: Solr
>  Issue Type: Bug
>  Components: web gui
>Affects Versions: 4.1
> Environment: Oracle Corporation OpenJDK 64-Bit Server VM (1.7.0_15 
> 23.7-b01)
>Reporter: Shawn Heisey
>Assignee: Stefan Matheis (steffkes)
> Attachments: built-in-jetty8.png, ubuntu-jetty6.png
>
>
> A user on IRC was having trouble getting file descriptor counts and JVM 
> memory usage in the admin UI, but it worked perfectly fine on another system. 
>  The problem system uses OpenJDK, the other one uses the Apple JDK.  The user 
> had tracked it down to an exception while trying to get open file descriptor 
> info.
> Looking in the SystemInfoHandler.java file, I see a comment reference to 
> com.sun.management.UnixOperatingSystemMXBean at the point where it is getting 
> file descriptor info.  A little extra searching turned up ZOOKEEPER-1579 
> which refers to HBASE-6945, the same problem with OpenJDK.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



RE: verbose logging

2013-03-04 Thread Uwe Schindler
Oh,

I though the XML report files contain the whole stdout?

http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/ws/solr/build/solr-core/test/TEST-org.apache.solr.update.TestAtomicUpdateErrorCases.xml

But its empty... This would be the most "correct" variant. I know that by 
default stock JUnit saves the whole stdout/stderr in the XML file, but JUnit4 
runner does not seem to do this?

Uwe

-
Uwe Schindler
H.-H.-Meier-Allee 63, D-28213 Bremen
http://www.thetaphi.de
eMail: u...@thetaphi.de


> -Original Message-
> From: dawid.we...@gmail.com [mailto:dawid.we...@gmail.com] On Behalf
> Of Dawid Weiss
> Sent: Monday, March 04, 2013 9:55 PM
> To: dev@lucene.apache.org
> Subject: Re: verbose logging
> 
> The system variable has nothing to do with it. The output from tests is passed
> to report plugins and the default one printing to console simply discards
> whatever goes to sysout/syserr if a test passed (and within its suite's
> boundary).  There is another report plugin that saves full output to disk so
> you have a few choices:
> 
> 1) enable full sysout/syserr logging by passing tests.showSuccess=true to
> jenkins (or committing it in to the repo),
> 2) going to a particular jenkins plan, accessing the workspace and fetching
> build output for that test. For example:
> http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-
> Linux/ws/lucene/build/core/test/tests-report.txt
> 
> (this file may not be there if a build hasn't reached it yet).
> 
> 3) ask Uwe or whoever is the jenkins admin to preserve **/tests-report.txt
> as a build artefact; this way you'd have some history to look at. These files
> are *big* though, especially for Solr which emits tons of logs.
> 
> Dawid
> 
> On Mon, Mar 4, 2013 at 9:47 PM, Erick Erickson 
> wrote:
> > I wasn't very clear. Since I can't get the tests to fail in my local
> > environment, I was trying for the verbose output for this particular
> > test _whenever_ the test was run. Is it reasonable to set the system
> > var for tests.showSuccess=true in my @BeforeClass?
> >
> > Thanks,
> > Erick
> >
> >
> > On Mon, Mar 4, 2013 at 9:52 AM, Dawid Weiss
> > 
> > wrote:
> >>
> >> By default the output from tests that passed is not dumped to the
> >> console (you wouldn't be able to track which JVM dumped what anyway
> >> when running on multicore). If you need a full log then:
> >>
> >> > ant test-help
> >> ...
> >>  [echo] # Include additional information like what is printed to
> >>  [echo] # sysout/syserr, even if the test passes.
> >>  [echo] # Enabled automatically when running for a single test case.
> >>  [echo] ant -Dtests.showSuccess=true test
> >>
> >> Should do the job. If you'd like to see the output from jenkins or
> >> any other build server then, unless it fails, it won't be on the console.
> >>
> >> Dawid
> >>
> >> On Mon, Mar 4, 2013 at 3:48 PM, Erick Erickson
> >> 
> >> wrote:
> >> > I tried turning on verbose in the OpenCloseCoreStressTests by
> >> > setting System.setProperty("tests.verbose", "true"); in
> >> > @BeforeClass and, of course, it works on my machine. But the failed
> >> > test logs don't have verbose output, is there some magic here? For
> >> > the nonce, I'm just doing a bit of hard logging.
> >> >
> >> > I also found another place where I screwed up the a Stream by not
> >> > closing it, ZkCLI of all places. I've checked that in.
> >> >
> >> > Of course all tests pass on my machine. Scant comfort that.
> >> >
> >> > I also had yet another assumption proved wrong so I didn't get any
> >> > useful info from last night's failure, trying to add some more
> >> > logging again today.
> >> > Don't panic that the checking has my initials init, that's so I'm
> >> > absolutely sure to know to take it out.
> >> >
> >> > Thanks,
> >> > Erick
> >>
> >> -
> >> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For
> >> additional commands, e-mail: dev-h...@lucene.apache.org
> >>
> >
> 
> -
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional
> commands, e-mail: dev-h...@lucene.apache.org


-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-4138) currency field doesn't work with functions (ie: isn't compatible with frange query)

2013-03-04 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4138?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man resolved SOLR-4138.


   Resolution: Fixed
Fix Version/s: 5.0
   4.2

Committed revision 1452483.
Committed revision 1452508.
Committed revision 1452527.


> currency field doesn't work with functions (ie: isn't compatible with frange 
> query)
> ---
>
> Key: SOLR-4138
> URL: https://issues.apache.org/jira/browse/SOLR-4138
> Project: Solr
>  Issue Type: Bug
>  Components: search
>Affects Versions: 4.0
>Reporter: Grzegorz Sobczyk
>Assignee: Hoss Man
> Fix For: 4.2, 5.0
>
> Attachments: SOLR-4135-test.patch, SOLR-4138.patch, SOLR-4138.patch
>
>
> In general, using CurrencyField with FunctionQueries doesn't work
> In particular, as originally reported...
> Filtering using {!frange} syntax isn't work properly. (rather at all)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Assigned] (SOLR-4526) Admin UI dependence on UnixOperationSystemMXBean

2013-03-04 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4526?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man reassigned SOLR-4526:
--

Assignee: Stefan Matheis (steffkes)

Felix: can you confirm no errors in the server logs, just javacsript was having 
problems from the null keys?

asigning to steffkes in the hopes that he can take a look -- but based on the 
screen shot i'm not really sure what if anything it could do better in a 
situation like this -- if the values aren't available from the JVM, showing 
"blank" bar sharts seems as good an option as any ... i guess maybe labeling 
them with "?" could be a little better though.

> Admin UI dependence on UnixOperationSystemMXBean
> 
>
> Key: SOLR-4526
> URL: https://issues.apache.org/jira/browse/SOLR-4526
> Project: Solr
>  Issue Type: Bug
>  Components: web gui
>Affects Versions: 4.1
> Environment: Oracle Corporation OpenJDK 64-Bit Server VM (1.7.0_15 
> 23.7-b01)
>Reporter: Shawn Heisey
>Assignee: Stefan Matheis (steffkes)
> Attachments: built-in-jetty8.png, ubuntu-jetty6.png
>
>
> A user on IRC was having trouble getting file descriptor counts and JVM 
> memory usage in the admin UI, but it worked perfectly fine on another system. 
>  The problem system uses OpenJDK, the other one uses the Apple JDK.  The user 
> had tracked it down to an exception while trying to get open file descriptor 
> info.
> Looking in the SystemInfoHandler.java file, I see a comment reference to 
> com.sun.management.UnixOperatingSystemMXBean at the point where it is getting 
> file descriptor info.  A little extra searching turned up ZOOKEEPER-1579 
> which refers to HBASE-6945, the same problem with OpenJDK.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Ability to specify 2 different query analyzers for same indexed field in Solr

2013-03-04 Thread Tom Burton-West
Hello,

We would like to be able to specify two different fields that both use the
same indexed field but use different analyzers.   An example use-case for
this might be doing query-time synonym expansion with the synonyms weighted
lower than an exact match.

q=exact_field^10 OR synonyms^1

The normal way to do this in Solr, which is just to set up separate
analyzer chains and use a copyfield, will not work for us because the field
in question is huge.  It is about 7 TB of OCR.

Is there a way to do this currently in Solr?   If not ,

1) should I open a JIRA issue?
2) can someone point me towards the part of the code I might need to modify?

Tom

Tom Burton-West
Information Retrieval Programmer
Digital Library Production Service
University of Michigan Library
http://www.hathitrust.org/blogs/large-scale-search


[jira] [Updated] (SOLR-4526) Admin UI dependence on UnixOperationSystemMXBean

2013-03-04 Thread Felix Buenemann (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4526?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Felix Buenemann updated SOLR-4526:
--

Attachment: built-in-jetty8.png
ubuntu-jetty6.png

Screenshots illustrating the problem. The JVM values are only empty because of 
an exception in index.js, which was looking for the missing keys 
'maxFileDescriptorCount' and 'openFileDescriptorCount' in system_data['system'].

> Admin UI dependence on UnixOperationSystemMXBean
> 
>
> Key: SOLR-4526
> URL: https://issues.apache.org/jira/browse/SOLR-4526
> Project: Solr
>  Issue Type: Bug
>  Components: web gui
>Affects Versions: 4.1
> Environment: Oracle Corporation OpenJDK 64-Bit Server VM (1.7.0_15 
> 23.7-b01)
>Reporter: Shawn Heisey
> Attachments: built-in-jetty8.png, ubuntu-jetty6.png
>
>
> A user on IRC was having trouble getting file descriptor counts and JVM 
> memory usage in the admin UI, but it worked perfectly fine on another system. 
>  The problem system uses OpenJDK, the other one uses the Apple JDK.  The user 
> had tracked it down to an exception while trying to get open file descriptor 
> info.
> Looking in the SystemInfoHandler.java file, I see a comment reference to 
> com.sun.management.UnixOperatingSystemMXBean at the point where it is getting 
> file descriptor info.  A little extra searching turned up ZOOKEEPER-1579 
> which refers to HBASE-6945, the same problem with OpenJDK.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: verbose logging

2013-03-04 Thread Dawid Weiss
The system variable has nothing to do with it. The output from tests
is passed to report plugins and the default one printing to console
simply discards whatever goes to sysout/syserr if a test passed (and
within its suite's boundary).  There is another report plugin that
saves full output to disk so you have a few choices:

1) enable full sysout/syserr logging by passing tests.showSuccess=true
to jenkins (or committing it in to the repo),
2) going to a particular jenkins plan, accessing the workspace and
fetching build output for that test. For example:
http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-Linux/ws/lucene/build/core/test/tests-report.txt

(this file may not be there if a build hasn't reached it yet).

3) ask Uwe or whoever is the jenkins admin to preserve
**/tests-report.txt as a build artefact; this way you'd have some
history to look at. These files are *big* though, especially for Solr
which emits tons of logs.

Dawid

On Mon, Mar 4, 2013 at 9:47 PM, Erick Erickson  wrote:
> I wasn't very clear. Since I can't get the tests to fail in my local
> environment, I was trying for the verbose output for this particular test
> _whenever_ the test was run. Is it reasonable to set the system var for
> tests.showSuccess=true in my @BeforeClass?
>
> Thanks,
> Erick
>
>
> On Mon, Mar 4, 2013 at 9:52 AM, Dawid Weiss 
> wrote:
>>
>> By default the output from tests that passed is not dumped to the
>> console (you wouldn't be able to track which JVM dumped what anyway
>> when running on multicore). If you need a full log then:
>>
>> > ant test-help
>> ...
>>  [echo] # Include additional information like what is printed to
>>  [echo] # sysout/syserr, even if the test passes.
>>  [echo] # Enabled automatically when running for a single test case.
>>  [echo] ant -Dtests.showSuccess=true test
>>
>> Should do the job. If you'd like to see the output from jenkins or any
>> other build server then, unless it fails, it won't be on the console.
>>
>> Dawid
>>
>> On Mon, Mar 4, 2013 at 3:48 PM, Erick Erickson 
>> wrote:
>> > I tried turning on verbose in the OpenCloseCoreStressTests by setting
>> > System.setProperty("tests.verbose", "true");
>> > in @BeforeClass and, of course, it works on my machine. But the failed
>> > test
>> > logs don't have verbose output, is there some magic here? For the nonce,
>> > I'm
>> > just doing a bit of hard logging.
>> >
>> > I also found another place where I screwed up the a Stream by not
>> > closing
>> > it, ZkCLI of all places. I've checked that in.
>> >
>> > Of course all tests pass on my machine. Scant comfort that.
>> >
>> > I also had yet another assumption proved wrong so I didn't get any
>> > useful
>> > info from last night's failure, trying to add some more logging again
>> > today.
>> > Don't panic that the checking has my initials init, that's so I'm
>> > absolutely
>> > sure to know to take it out.
>> >
>> > Thanks,
>> > Erick
>>
>> -
>> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
>> For additional commands, e-mail: dev-h...@lucene.apache.org
>>
>

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4416) Upgrade to Tika 1.3

2013-03-04 Thread Commit Tag Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4416?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592624#comment-13592624
 ] 

Commit Tag Bot commented on SOLR-4416:
--

[branch_4x commit] Mark Robert Miller
http://svn.apache.org/viewvc?view=revision&revision=1452504

SOLR-4416: Upgrade to Tika 1.3.


> Upgrade to Tika 1.3
> ---
>
> Key: SOLR-4416
> URL: https://issues.apache.org/jira/browse/SOLR-4416
> Project: Solr
>  Issue Type: Task
>  Components: contrib - Solr Cell (Tika extraction)
>Affects Versions: 4.1
>Reporter: Erik Hatcher
>Assignee: Mark Miller
>Priority: Critical
> Fix For: 4.2, 5.0
>
> Attachments: SOLR-4416-trunk-1.patch
>
>
> Tika 1.3 was recently released with these improvements: 
> http://www.apache.org/dist/tika/CHANGES-1.3.txt

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-4416) Upgrade to Tika 1.3

2013-03-04 Thread Mark Miller (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4416?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Miller resolved SOLR-4416.
---

Resolution: Fixed

Committed to 4x and 5x - not sure why the 4.2 commit msg is so slow...

Thanks Markus!

> Upgrade to Tika 1.3
> ---
>
> Key: SOLR-4416
> URL: https://issues.apache.org/jira/browse/SOLR-4416
> Project: Solr
>  Issue Type: Task
>  Components: contrib - Solr Cell (Tika extraction)
>Affects Versions: 4.1
>Reporter: Erik Hatcher
>Assignee: Mark Miller
>Priority: Critical
> Fix For: 4.2, 5.0
>
> Attachments: SOLR-4416-trunk-1.patch
>
>
> Tika 1.3 was recently released with these improvements: 
> http://www.apache.org/dist/tika/CHANGES-1.3.txt

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4526) Admin UI dependence on UnixOperationSystemMXBean

2013-03-04 Thread Felix Buenemann (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4526?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592616#comment-13592616
 ] 

Felix Buenemann commented on SOLR-4526:
---

The problem turned out to be jetty6 vs. jetty8 and not OpenJDK. The dev system 
was running Solr's built-in jetty, while the production system was running the 
jetty version that was shipping with Ubuntu 12.04 Server. I noticed this after 
building a VM to simulate the problem with the closed Oracle JDK7 and 
encountered the same behavior.

I'm attaching two screenshots to show the problem.

> Admin UI dependence on UnixOperationSystemMXBean
> 
>
> Key: SOLR-4526
> URL: https://issues.apache.org/jira/browse/SOLR-4526
> Project: Solr
>  Issue Type: Bug
>  Components: web gui
>Affects Versions: 4.1
> Environment: Oracle Corporation OpenJDK 64-Bit Server VM (1.7.0_15 
> 23.7-b01)
>Reporter: Shawn Heisey
>
> A user on IRC was having trouble getting file descriptor counts and JVM 
> memory usage in the admin UI, but it worked perfectly fine on another system. 
>  The problem system uses OpenJDK, the other one uses the Apple JDK.  The user 
> had tracked it down to an exception while trying to get open file descriptor 
> info.
> Looking in the SystemInfoHandler.java file, I see a comment reference to 
> com.sun.management.UnixOperatingSystemMXBean at the point where it is getting 
> file descriptor info.  A little extra searching turned up ZOOKEEPER-1579 
> which refers to HBASE-6945, the same problem with OpenJDK.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: verbose logging

2013-03-04 Thread Uwe Schindler
Jenkins saves complete sysout. Jou just have to go to build artifacts and 
download the test xml file.



Erick Erickson  schrieb:

>I wasn't very clear. Since I can't get the tests to fail in my local
>environment, I was trying for the verbose output for this particular
>test
>_whenever_ the test was run. Is it reasonable to set the system var for
>tests.showSuccess=true in my @BeforeClass?
>
>Thanks,
>Erick
>
>
>On Mon, Mar 4, 2013 at 9:52 AM, Dawid Weiss
>wrote:
>
>> By default the output from tests that passed is not dumped to the
>> console (you wouldn't be able to track which JVM dumped what anyway
>> when running on multicore). If you need a full log then:
>>
>> > ant test-help
>> ...
>>  [echo] # Include additional information like what is printed to
>>  [echo] # sysout/syserr, even if the test passes.
>>  [echo] # Enabled automatically when running for a single test
>case.
>>  [echo] ant -Dtests.showSuccess=true test
>>
>> Should do the job. If you'd like to see the output from jenkins or
>any
>> other build server then, unless it fails, it won't be on the console.
>>
>> Dawid
>>
>> On Mon, Mar 4, 2013 at 3:48 PM, Erick Erickson
>
>> wrote:
>> > I tried turning on verbose in the OpenCloseCoreStressTests by
>setting
>> > System.setProperty("tests.verbose", "true");
>> > in @BeforeClass and, of course, it works on my machine. But the
>failed
>> test
>> > logs don't have verbose output, is there some magic here? For the
>nonce,
>> I'm
>> > just doing a bit of hard logging.
>> >
>> > I also found another place where I screwed up the a Stream by not
>closing
>> > it, ZkCLI of all places. I've checked that in.
>> >
>> > Of course all tests pass on my machine. Scant comfort that.
>> >
>> > I also had yet another assumption proved wrong so I didn't get any
>useful
>> > info from last night's failure, trying to add some more logging
>again
>> today.
>> > Don't panic that the checking has my initials init, that's so I'm
>> absolutely
>> > sure to know to take it out.
>> >
>> > Thanks,
>> > Erick
>>
>> -
>> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
>> For additional commands, e-mail: dev-h...@lucene.apache.org
>>
>>

--
Uwe Schindler
H.-H.-Meier-Allee 63, 28213 Bremen
http://www.thetaphi.de

Re: verbose logging

2013-03-04 Thread Erick Erickson
I wasn't very clear. Since I can't get the tests to fail in my local
environment, I was trying for the verbose output for this particular test
_whenever_ the test was run. Is it reasonable to set the system var for
tests.showSuccess=true in my @BeforeClass?

Thanks,
Erick


On Mon, Mar 4, 2013 at 9:52 AM, Dawid Weiss wrote:

> By default the output from tests that passed is not dumped to the
> console (you wouldn't be able to track which JVM dumped what anyway
> when running on multicore). If you need a full log then:
>
> > ant test-help
> ...
>  [echo] # Include additional information like what is printed to
>  [echo] # sysout/syserr, even if the test passes.
>  [echo] # Enabled automatically when running for a single test case.
>  [echo] ant -Dtests.showSuccess=true test
>
> Should do the job. If you'd like to see the output from jenkins or any
> other build server then, unless it fails, it won't be on the console.
>
> Dawid
>
> On Mon, Mar 4, 2013 at 3:48 PM, Erick Erickson 
> wrote:
> > I tried turning on verbose in the OpenCloseCoreStressTests by setting
> > System.setProperty("tests.verbose", "true");
> > in @BeforeClass and, of course, it works on my machine. But the failed
> test
> > logs don't have verbose output, is there some magic here? For the nonce,
> I'm
> > just doing a bit of hard logging.
> >
> > I also found another place where I screwed up the a Stream by not closing
> > it, ZkCLI of all places. I've checked that in.
> >
> > Of course all tests pass on my machine. Scant comfort that.
> >
> > I also had yet another assumption proved wrong so I didn't get any useful
> > info from last night's failure, trying to add some more logging again
> today.
> > Don't panic that the checking has my initials init, that's so I'm
> absolutely
> > sure to know to take it out.
> >
> > Thanks,
> > Erick
>
> -
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
> For additional commands, e-mail: dev-h...@lucene.apache.org
>
>


Re: svn commit: r1452483 - in /lucene/dev/trunk/solr: ./ core/src/java/org/apache/solr/schema/ core/src/java/org/apache/solr/search/ core/src/test-files/solr/collection1/conf/ core/src/test/org/apache

2013-03-04 Thread Chris Hostetter

Bah ... i swear i ran "ant precommit" on this, and yet i just found 
doclint problems (merging to 4x) ... fix coming.

: Date: Mon, 04 Mar 2013 20:25:37 -
: From: hoss...@apache.org
: Reply-To: dev@lucene.apache.org
: To: comm...@lucene.apache.org
: Subject: svn commit: r1452483 - in /lucene/dev/trunk/solr: ./
: core/src/java/org/apache/solr/schema/
: core/src/java/org/apache/solr/search/
: core/src/test-files/solr/collection1/conf/
: core/src/test/org/apache/solr/schema/
: core/src/test/org/apache/solr/search/
: 
: Author: hossman
: Date: Mon Mar  4 20:25:37 2013
: New Revision: 1452483
: 
: URL: http://svn.apache.org/r1452483
: Log:
: SOLR-4138: CurrencyField fields can now be used in a ValueSources. There is 
also a new currency(field,[CODE]) function
: 
: Modified:
: lucene/dev/trunk/solr/CHANGES.txt
: 
lucene/dev/trunk/solr/core/src/java/org/apache/solr/schema/CurrencyField.java
: 
lucene/dev/trunk/solr/core/src/java/org/apache/solr/search/ValueSourceParser.java
: 
lucene/dev/trunk/solr/core/src/test-files/solr/collection1/conf/schema15.xml
: 
lucene/dev/trunk/solr/core/src/test/org/apache/solr/schema/AbstractCurrencyFieldTest.java
: 
lucene/dev/trunk/solr/core/src/test/org/apache/solr/search/QueryEqualityTest.java
: 
: Modified: lucene/dev/trunk/solr/CHANGES.txt
: URL: 
http://svn.apache.org/viewvc/lucene/dev/trunk/solr/CHANGES.txt?rev=1452483&r1=1452482&r2=1452483&view=diff
: ==
: --- lucene/dev/trunk/solr/CHANGES.txt (original)
: +++ lucene/dev/trunk/solr/CHANGES.txt Mon Mar  4 20:25:37 2013
: @@ -98,6 +98,14 @@ New Features
:Similarity when you know the optimal "Sweet Spot" of values for the field 
:length and TF scoring factors.  (hossman)
:  
: +* SOLR-4138: CurrencyField fields can now be used in a ValueSources to
: +  get the "raw" value (using the default number of fractional digits) in 
: +  the default currency of the field type.  There is also a new 
: +  currency(field,[CODE]) function for generating a ValueSource of the 
: +  "natural" value, converted to an optionally specified currency to 
: +  override the default for the field type.
: +  (hossman)
: +
:  Bug Fixes
:  --
:  
: 
: Modified: 
lucene/dev/trunk/solr/core/src/java/org/apache/solr/schema/CurrencyField.java
: URL: 
http://svn.apache.org/viewvc/lucene/dev/trunk/solr/core/src/java/org/apache/solr/schema/CurrencyField.java?rev=1452483&r1=1452482&r2=1452483&view=diff
: ==
: --- 
lucene/dev/trunk/solr/core/src/java/org/apache/solr/schema/CurrencyField.java 
(original)
: +++ 
lucene/dev/trunk/solr/core/src/java/org/apache/solr/schema/CurrencyField.java 
Mon Mar  4 20:25:37 2013
: @@ -242,6 +242,67 @@ public class CurrencyField extends Field
:  return getRangeQuery(parser, field, valueDefault, valueDefault, true, 
true);
:}
:  
: +  /**
: +   * 
: +   * Returns a ValueSource over this field in which the numeric value for 
: +   * each document represents the indexed value as converted to the default 
: +   * currency for the field, normalized to it's most granular form based 
: +   * on the default fractional digits.
: +   * 
: +   * 
: +   * For example: If the default Currency specified for a field is 
: +   * USD, then the values returned by this value source would 
: +   * represent the equivilent number of "cents" (ie: value in dollars * 100) 
: +   * after converting each document's native currency to USD -- because the 
: +   * default fractional digits for USD is "2".  
: +   * So for a document whose indexed value was currently equivilent to 
: +   * "5.43,USD" using the the exchange provider for this field, 
: +   * this ValueSource would return a value of "543"
: +   * 
: +   *
: +   * @see #PARAM_DEFAULT_CURRENCY
: +   * @see #DEFAULT_DEFAULT_CURRENCY
: +   * @see Currency#getDefaultFractionDigits
: +   * @see getConvertedValueSource
: +   */
: +  public RawCurrencyValueSource getValueSource(SchemaField field, 
: +   QParser parser) {
: +field.checkFieldCacheSource(parser);
: +return new RawCurrencyValueSource(field, defaultCurrency, parser);
: +  }
: +
: +  /**
: +   * 
: +   * Returns a ValueSource over this field in which the numeric value for 
: +   * each document represents the value from the underlying 
: +   * RawCurrencyValueSource as converted to the specified 
target 
: +   * Currency.
: +   * 
: +   * 
: +   * For example: If the targetCurrencyCode param is set to
: +   * USD, then the values returned by this value source would 
: +   * represent the equivilent number of dollars after converting each 
: +   * document's raw value to USD.  So for a document whose 
: +   * indexed value was currently equivilent to "5.43,USD" 
: +   * using the the exchange provider for this field, this ValueSource would 
: +   * return a value

Re: Lucene/Solr 4.2

2013-03-04 Thread Walter Underwood
I'd appreciate it if this was in:

https://issues.apache.org/jira/browse/LUCENE-4810

wunder

On Mar 4, 2013, at 11:58 AM, Mark Miller wrote:

> Just a reminder - I'm looking to start pushing this release very soon. I'd 
> like to get an RC up by early next week if I can. 
> 
> Please help me out by getting what you want into 4.2 as soon as possible.
> 
> - Mark
> -
> To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
> For additional commands, e-mail: dev-h...@lucene.apache.org
> 






[jira] [Commented] (SOLR-4416) Upgrade to Tika 1.3

2013-03-04 Thread Commit Tag Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4416?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592611#comment-13592611
 ] 

Commit Tag Bot commented on SOLR-4416:
--

[trunk commit] Mark Robert Miller
http://svn.apache.org/viewvc?view=revision&revision=1452499

SOLR-4416: Upgrade to Tika 1.3.


> Upgrade to Tika 1.3
> ---
>
> Key: SOLR-4416
> URL: https://issues.apache.org/jira/browse/SOLR-4416
> Project: Solr
>  Issue Type: Task
>  Components: contrib - Solr Cell (Tika extraction)
>Affects Versions: 4.1
>Reporter: Erik Hatcher
>Assignee: Mark Miller
>Priority: Critical
> Fix For: 4.2, 5.0
>
> Attachments: SOLR-4416-trunk-1.patch
>
>
> Tika 1.3 was recently released with these improvements: 
> http://www.apache.org/dist/tika/CHANGES-1.3.txt

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4138) currency field doesn't work with functions (ie: isn't compatible with frange query)

2013-03-04 Thread Commit Tag Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4138?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592603#comment-13592603
 ] 

Commit Tag Bot commented on SOLR-4138:
--

[trunk commit] Chris M. Hostetter
http://svn.apache.org/viewvc?view=revision&revision=1452483

SOLR-4138: CurrencyField fields can now be used in a ValueSources. There is 
also a new currency(field,[CODE]) function


> currency field doesn't work with functions (ie: isn't compatible with frange 
> query)
> ---
>
> Key: SOLR-4138
> URL: https://issues.apache.org/jira/browse/SOLR-4138
> Project: Solr
>  Issue Type: Bug
>  Components: search
>Affects Versions: 4.0
>Reporter: Grzegorz Sobczyk
>Assignee: Hoss Man
> Attachments: SOLR-4135-test.patch, SOLR-4138.patch, SOLR-4138.patch
>
>
> In general, using CurrencyField with FunctionQueries doesn't work
> In particular, as originally reported...
> Filtering using {!frange} syntax isn't work properly. (rather at all)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4526) Admin UI dependence on UnixOperationSystemMXBean

2013-03-04 Thread Shawn Heisey (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4526?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592597#comment-13592597
 ] 

Shawn Heisey commented on SOLR-4526:


Additional note from the user on IRC: I adjusted the live server config to run 
the solr built-in jetty and now all the stats are showing up, even on openjdk, 
so the issue seems to be with older jetty versions.

> Admin UI dependence on UnixOperationSystemMXBean
> 
>
> Key: SOLR-4526
> URL: https://issues.apache.org/jira/browse/SOLR-4526
> Project: Solr
>  Issue Type: Bug
>  Components: web gui
>Affects Versions: 4.1
> Environment: Oracle Corporation OpenJDK 64-Bit Server VM (1.7.0_15 
> 23.7-b01)
>Reporter: Shawn Heisey
>
> A user on IRC was having trouble getting file descriptor counts and JVM 
> memory usage in the admin UI, but it worked perfectly fine on another system. 
>  The problem system uses OpenJDK, the other one uses the Apple JDK.  The user 
> had tracked it down to an exception while trying to get open file descriptor 
> info.
> Looking in the SystemInfoHandler.java file, I see a comment reference to 
> com.sun.management.UnixOperatingSystemMXBean at the point where it is getting 
> file descriptor info.  A little extra searching turned up ZOOKEEPER-1579 
> which refers to HBASE-6945, the same problem with OpenJDK.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Assigned] (SOLR-4416) Upgrade to Tika 1.3

2013-03-04 Thread Mark Miller (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4416?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Miller reassigned SOLR-4416:
-

Assignee: Mark Miller

> Upgrade to Tika 1.3
> ---
>
> Key: SOLR-4416
> URL: https://issues.apache.org/jira/browse/SOLR-4416
> Project: Solr
>  Issue Type: Task
>  Components: contrib - Solr Cell (Tika extraction)
>Affects Versions: 4.1
>Reporter: Erik Hatcher
>Assignee: Mark Miller
>Priority: Critical
> Fix For: 4.2, 5.0
>
> Attachments: SOLR-4416-trunk-1.patch
>
>
> Tika 1.3 was recently released with these improvements: 
> http://www.apache.org/dist/tika/CHANGES-1.3.txt

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4811) *DocValuesField should extend Field not StoredField

2013-03-04 Thread Commit Tag Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4811?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592582#comment-13592582
 ] 

Commit Tag Bot commented on LUCENE-4811:


[trunk commit] Michael McCandless
http://svn.apache.org/viewvc?view=revision&revision=1452474

LUCENE-4811: fix *DVField to extend Field not StoredField


> *DocValuesField should extend Field not StoredField
> ---
>
> Key: LUCENE-4811
> URL: https://issues.apache.org/jira/browse/LUCENE-4811
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Michael McCandless
>Assignee: Michael McCandless
> Fix For: 5.0
>
> Attachments: LUCENE-4811.patch
>
>
> I think it's confusing because it implies the field will be stored as well as 
> added to doc values ...
> I also noticed StorableFieldType is unused.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4526) Admin UI dependence on UnixOperationSystemMXBean

2013-03-04 Thread Shawn Heisey (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4526?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592581#comment-13592581
 ] 

Shawn Heisey commented on SOLR-4526:


I think perhaps I was a little too aggressive in framing this issue.  I should 
have scaled it back a little bit to focus on the UI issue, not the java 
classes.  The title may need changing, and it may require filing another issue. 
 I see two problems.  With OpenJDK, the file descriptor info and the JVM memory 
info are not showing on the dashboard.  Side note: I don't have OpenJDK running 
anywhere and don't have unallocated hardware I can fiddle with right now.

The underlying issue is that the user says the file descriptor info is not 
found in /admin/system output running under OpenJDK, but it is there under 
Apple or Oracle JVMs.  If the info is available at all from OpenJDK, it may 
require a code change to get it.

The java issue is triggering a javascript issue.  An educated guess is that the 
code that populates the file descriptor bar is failing and causing the JVM 
memory bar calculation (JVM memory info is available in /admin/system) to be 
skipped entirely.  A simple fix for that issue would be to re-order the 
javascript code so that jvm memory is done before the file descriptor.  A 
better fix would be to make sure that a nonexistent piece of information 
doesn't cause problems.

I'm very weak in both javascript and java reflection, or I would have already 
tried to come up with patches.


> Admin UI dependence on UnixOperationSystemMXBean
> 
>
> Key: SOLR-4526
> URL: https://issues.apache.org/jira/browse/SOLR-4526
> Project: Solr
>  Issue Type: Bug
>  Components: web gui
>Affects Versions: 4.1
> Environment: Oracle Corporation OpenJDK 64-Bit Server VM (1.7.0_15 
> 23.7-b01)
>Reporter: Shawn Heisey
>
> A user on IRC was having trouble getting file descriptor counts and JVM 
> memory usage in the admin UI, but it worked perfectly fine on another system. 
>  The problem system uses OpenJDK, the other one uses the Apple JDK.  The user 
> had tracked it down to an exception while trying to get open file descriptor 
> info.
> Looking in the SystemInfoHandler.java file, I see a comment reference to 
> com.sun.management.UnixOperatingSystemMXBean at the point where it is getting 
> file descriptor info.  A little extra searching turned up ZOOKEEPER-1579 
> which refers to HBASE-6945, the same problem with OpenJDK.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-4811) *DocValuesField should extend Field not StoredField

2013-03-04 Thread Michael McCandless (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4811?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Michael McCandless resolved LUCENE-4811.


Resolution: Fixed

> *DocValuesField should extend Field not StoredField
> ---
>
> Key: LUCENE-4811
> URL: https://issues.apache.org/jira/browse/LUCENE-4811
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Michael McCandless
>Assignee: Michael McCandless
> Fix For: 5.0
>
> Attachments: LUCENE-4811.patch
>
>
> I think it's confusing because it implies the field will be stored as well as 
> added to doc values ...
> I also noticed StorableFieldType is unused.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4465) Configurable Collectors

2013-03-04 Thread Joel Bernstein (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4465?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592576#comment-13592576
 ] 

Joel Bernstein commented on SOLR-4465:
--

Dan,

This patch doesn't work with grouping because grouping is done within a 
separate flow with a separate set of collectors. SOLR-1680 was a custom 
collectors patch that was meant to allow grouping to be merged into the main 
flow, but was never committed. Possibly grouping could be implemented using a 
collector factory but that would be a follow-on to getting this as part of the 
core.

I'll work on resolving the npe with the default collector in the next patch.

As you mentioned the caching implications here need to be fully understood. 
Perhaps you could elaborate on "Additionally since the collector is free to 
alter results between requests". Can you explain the use case where this would 
occur?



> Configurable Collectors
> ---
>
> Key: SOLR-4465
> URL: https://issues.apache.org/jira/browse/SOLR-4465
> Project: Solr
>  Issue Type: New Feature
>  Components: search
>Affects Versions: 4.1
>Reporter: Joel Bernstein
> Fix For: 4.2, 5.0
>
> Attachments: SOLR-4465.patch, SOLR-4465.patch, SOLR-4465.patch, 
> SOLR-4465.patch, SOLR-4465.patch, SOLR-4465.patch, SOLR-4465.patch, 
> SOLR-4465.patch, SOLR-4465.patch
>
>
> This issue is to add configurable custom collectors to Solr. This expands the 
> design and work done in issue SOLR-1680 to include:
> 1) CollectorFactory configuration in solconfig.xml
> 2) Http parameters to allow clients to dynamically select a CollectorFactory 
> and construct a custom Collector.
> 3) Make aspects of QueryComponent pluggable so that the output from 
> distributed search can conform with custom collectors at the shard level.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Lucene/Solr 4.2

2013-03-04 Thread Mark Miller
Just a reminder - I'm looking to start pushing this release very soon. I'd like 
to get an RC up by early next week if I can. 

Please help me out by getting what you want into 4.2 as soon as possible.

- Mark
-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-4527) Atomic updates when running distributed seem broken.

2013-03-04 Thread Mark Miller (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4527?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mark Miller updated SOLR-4527:
--

Fix Version/s: 5.0
   4.2

> Atomic updates when running distributed seem broken.
> 
>
> Key: SOLR-4527
> URL: https://issues.apache.org/jira/browse/SOLR-4527
> Project: Solr
>  Issue Type: Bug
>  Components: SolrCloud, update
>Affects Versions: 4.1
>Reporter: mike st. john
> Fix For: 4.2, 5.0
>
>
> When using solrcloud as a nosql solution,  i've run into the issue where i've 
> sent some atomic updates and i'm receiving an error  "missing required 
> field:"  implying that this is an add instead of an update.  when i add 
> distrib=false to the url and send the doc to the index where it resides, the 
> update is applied.
> Possibly related...when i try and do a real time get for the id,  its 
> throwing an NPE
>  "trace":"java.lang.NullPointerException\n\tat 
> org.apache.solr.handler.component.RealTimeGetComponent.createSubRequests(RealTimeGetComponent.java:368)\n\tat
>  
> org.apache.solr.handler.component.RealTimeGetComponent.distributedProcess(RealTimeGetComponent.java:325)\n\tat
>  
> org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:244)\n\tat
>  
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)\n\tat
>  org.apache.solr.core.SolrCore.execute(SolrCore.java:1808)\n\tat 
> org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:583)\n\tat
>  
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:293)\n\tat
>  
> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:243)\n\tat
>  
> org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:210)\n\tat
>  
> org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:222)\n\tat
>  
> org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:123)\n\tat
>  
> org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:171)\n\tat
>  
> org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:99)\n\tat
>  
> org.apache.catalina.valves.AccessLogValve.invoke(AccessLogValve.java:936)\n\tat
>  
> org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:118)\n\tat
>  
> org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:407)\n\tat
>  
> org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1004)\n\tat
>  
> org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:589)\n\tat
>  
> org.apache.tomcat.util.net.JIoEndpoint$SocketProcessor.run(JIoEndpoint.java:310)\n\tat
>  
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110)\n\tat
>  
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603)\n\tat
>  java.lang.Thread.run(Thread.java:679)\n",
> "code":500}}
> the command will succeed , if i use the url the doc exists on and add 
> distrib=false to the end.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-4526) Admin UI dependence on UnixOperationSystemMXBean

2013-03-04 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4526?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man updated SOLR-4526:
---

Description: 
A user on IRC was having trouble getting file descriptor counts and JVM memory 
usage in the admin UI, but it worked perfectly fine on another system.  The 
problem system uses OpenJDK, the other one uses the Apple JDK.  The user had 
tracked it down to an exception while trying to get open file descriptor info.

Looking in the SystemInfoHandler.java file, I see a comment reference to 
com.sun.management.UnixOperatingSystemMXBean at the point where it is getting 
file descriptor info.  A little extra searching turned up ZOOKEEPER-1579 which 
refers to HBASE-6945, the same problem with OpenJDK.


  was:
A user on IRC was having trouble getting file descriptor counts and JVM memory 
usage in the admin UI, but it worked perfectly fine on another system.  The 
problem system uses OpenJDK, the other one uses the Apple JDK.  The user had 
tracked it down to an exception while trying to get open file descriptor info.

Looking in the SystemInfoHandler.java file, I see a comment reference to 
com.sun.management.UnixOperatingSystemMXBean at the point where it is getting 
file descriptor info.  A little extra searching turned up ZOOKEEPER-1579 which 
refers to HBASE-6495, the same problem with OpenJDK.



I'm confused: SystemInfoHandler doesn't directly depend on 
UnixOperatingSystemMXBean, it uses reflection to ensure that it only calls 
methods on the OperatingSystemMXBean if they already exist -- so even if it 
isn't a sun JVM, or isn't an UnixOperatingSystemMXBean instance, or is an older 
version of UnixOperatingSystemMXBean that doesn't have some of those methods, 
it still shouldn't be an error -- at worst those properties just won't be 
included in the response.

perhaps the problem is just that that admin UI Javascript assumes those 
properties will always be available? (but that wouldn't explain the comment 
about "tracked it down to an exception while trying to get open file descriptor 
info")

can anyone who can actually reproduce this problem please post the specifics of 
the exception they see, or the behavior they see in the admin ui (ie: 
screenshot)

> Admin UI dependence on UnixOperationSystemMXBean
> 
>
> Key: SOLR-4526
> URL: https://issues.apache.org/jira/browse/SOLR-4526
> Project: Solr
>  Issue Type: Bug
>  Components: web gui
>Affects Versions: 4.1
> Environment: Oracle Corporation OpenJDK 64-Bit Server VM (1.7.0_15 
> 23.7-b01)
>Reporter: Shawn Heisey
>
> A user on IRC was having trouble getting file descriptor counts and JVM 
> memory usage in the admin UI, but it worked perfectly fine on another system. 
>  The problem system uses OpenJDK, the other one uses the Apple JDK.  The user 
> had tracked it down to an exception while trying to get open file descriptor 
> info.
> Looking in the SystemInfoHandler.java file, I see a comment reference to 
> com.sun.management.UnixOperatingSystemMXBean at the point where it is getting 
> file descriptor info.  A little extra searching turned up ZOOKEEPER-1579 
> which refers to HBASE-6945, the same problem with OpenJDK.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4526) Admin UI dependence on UnixOperationSystemMXBean

2013-03-04 Thread Shawn Heisey (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4526?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592519#comment-13592519
 ] 

Shawn Heisey commented on SOLR-4526:


If File descriptor info just isn't available in OpenJDK, then that is life.  
The file descriptor problem seems to be causing another problem - the JVM 
memory info isn't showing up on the dashboard.  I haven't looked at the code 
yet, and I don't know javascript very well, but the user suspects that when the 
file descriptor info is not found, it causes the code that puts the JVM memory 
info together to be skipped.

> Admin UI dependence on UnixOperationSystemMXBean
> 
>
> Key: SOLR-4526
> URL: https://issues.apache.org/jira/browse/SOLR-4526
> Project: Solr
>  Issue Type: Bug
>  Components: web gui
>Affects Versions: 4.1
> Environment: Oracle Corporation OpenJDK 64-Bit Server VM (1.7.0_15 
> 23.7-b01)
>Reporter: Shawn Heisey
>
> A user on IRC was having trouble getting file descriptor counts and JVM 
> memory usage in the admin UI, but it worked perfectly fine on another system. 
>  The problem system uses OpenJDK, the other one uses the Apple JDK.  The user 
> had tracked it down to an exception while trying to get open file descriptor 
> info.
> Looking in the SystemInfoHandler.java file, I see a comment reference to 
> com.sun.management.UnixOperatingSystemMXBean at the point where it is getting 
> file descriptor info.  A little extra searching turned up ZOOKEEPER-1579 
> which refers to HBASE-6495, the same problem with OpenJDK.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-4527) Atomic updates when running distributed seem broken.

2013-03-04 Thread mike st. john (JIRA)
mike st. john created SOLR-4527:
---

 Summary: Atomic updates when running distributed seem broken.
 Key: SOLR-4527
 URL: https://issues.apache.org/jira/browse/SOLR-4527
 Project: Solr
  Issue Type: Bug
  Components: SolrCloud, update
Affects Versions: 4.1
Reporter: mike st. john


When using solrcloud as a nosql solution,  i've run into the issue where i've 
sent some atomic updates and i'm receiving an error  "missing required field:"  
implying that this is an add instead of an update.  when i add distrib=false to 
the url and send the doc to the index where it resides, the update is applied.

Possibly related...when i try and do a real time get for the id,  its throwing 
an NPE

 "trace":"java.lang.NullPointerException\n\tat 
org.apache.solr.handler.component.RealTimeGetComponent.createSubRequests(RealTimeGetComponent.java:368)\n\tat
 
org.apache.solr.handler.component.RealTimeGetComponent.distributedProcess(RealTimeGetComponent.java:325)\n\tat
 
org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:244)\n\tat
 
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)\n\tat
 org.apache.solr.core.SolrCore.execute(SolrCore.java:1808)\n\tat 
org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:583)\n\tat
 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:293)\n\tat
 
org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:243)\n\tat
 
org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:210)\n\tat
 
org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:222)\n\tat
 
org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:123)\n\tat
 
org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:171)\n\tat
 
org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:99)\n\tat
 
org.apache.catalina.valves.AccessLogValve.invoke(AccessLogValve.java:936)\n\tat 
org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:118)\n\tat
 
org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:407)\n\tat
 
org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1004)\n\tat
 
org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:589)\n\tat
 
org.apache.tomcat.util.net.JIoEndpoint$SocketProcessor.run(JIoEndpoint.java:310)\n\tat
 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110)\n\tat
 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603)\n\tat
 java.lang.Thread.run(Thread.java:679)\n",
"code":500}}


the command will succeed , if i use the url the doc exists on and add 
distrib=false to the end.



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Reopened] (SOLR-4524) ReturnFields could not support hash '#' in filed name

2013-03-04 Thread Hoss Man (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4524?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hoss Man reopened SOLR-4524:


  Assignee: Hoss Man

There is no problem here, i think there is just user confusion about how to 
include a "#" character in a URL -- "#" is a reserved character in URLs for 
dealing with relative anchors, when you type a URL like this in your browser...

{noformat}
http://localhost:8983/solr/select?q=id:HOSS&fl=score&fl=#foo_s&fl=id
{noformat}

...the browser never sends anything after the "#" to the server.

If you URL escape it, then everything works fine...

{noformat}
$ curl -sS 
'http://localhost:8983/solr/select?q=id:HOSS&fl=score&fl=%23foo_s&fl=id&indent=true&wt=json'
{
  "responseHeader":{
"status":0,
"QTime":1,
"params":{
  "fl":["score",
"#foo_s",
"id"],
  "indent":"true",
  "q":"id:HOSS",
  "wt":"json"}},
  "response":{"numFound":1,"start":0,"maxScore":3.8033605,"docs":[
  {
"id":"HOSS",
"#foo_s":"special",
"score":3.8033605}]
  }}
{noformat}

I've re-opened and assigned to myself to add a unit test demonstrating that a 
field name like this works in the fl param. 



> ReturnFields could not support hash '#' in filed name 
> --
>
> Key: SOLR-4524
> URL: https://issues.apache.org/jira/browse/SOLR-4524
> Project: Solr
>  Issue Type: Bug
>  Components: search
>Affects Versions: 4.1
>Reporter: Benjamin POUSSIN
>Assignee: Hoss Man
>  Labels: field, fl, query, queryparser
>
> If some fields start with '#' and we try to put it in fl param, this field 
> and all next fields are not in response document
> With query:
>   q=*:*&fl=score,#id,myOtherField
> Document response is like:
>SolrDocument{score=3.809621}
> this bug is similar to SOLR-2719 and SOLR-4374

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-4503) Add REST API methods to get schema information: fields, dynamicFields, fieldTypes, and copyFields

2013-03-04 Thread Steve Rowe (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4503?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Rowe updated SOLR-4503:
-

Fix Version/s: 4.2
  Description: 
Add REST methods that provide properties for fields, dynamicFields, fieldTypes, 
and copyFields, using paths:

/solr/(corename)/schema/fields
/solr/(corename)/schema/fields/fieldname

/solr/(corename)/schema/dynamicfields
/solr/(corename)/schema/dynamicfields/pattern

/solr/(corename)/schema/fieldtypes
/solr/(corename)/schema/fieldtypes/typename 

/solr/(corename)/schema/copyfields

  was:
Add REST methods that provide properties for fields, dynamic fields, and field 
types, using paths:

/solr/(corename)/schema/fields
/solr/(corename)/schema/fields/fieldname

/solr/(corename)/schema/dynamicfields
/solr/(corename)/schema/dynamicfields/pattern

/solr/(corename)/schema/fieldtypes
/solr/(corename)/schema/fieldtypes/typename 

  Summary: Add REST API methods to get schema information: fields, 
dynamicFields, fieldTypes, and copyFields  (was: Add REST API methods to get 
schema information: fields, dynamic fields, and field types)

> Add REST API methods to get schema information: fields, dynamicFields, 
> fieldTypes, and copyFields
> -
>
> Key: SOLR-4503
> URL: https://issues.apache.org/jira/browse/SOLR-4503
> Project: Solr
>  Issue Type: Sub-task
>  Components: Schema and Analysis
>Affects Versions: 4.1
>Reporter: Steve Rowe
>Assignee: Steve Rowe
> Fix For: 4.2
>
> Attachments: all.dynamic.fields.json, all.dynamic.fields.json, 
> all.fields.json, all.fields.json, all.field.types.json, all.field.types.json, 
> coordinate.dynamic.field.json, coordinate.dynamic.field.json, 
> copyfields.json, date.field.type.json, date.field.type.json, 
> price.field.json, price.field.json, SOLR-4503.patch, SOLR-4503.patch, 
> SOLR-4503.patch
>
>
> Add REST methods that provide properties for fields, dynamicFields, 
> fieldTypes, and copyFields, using paths:
> /solr/(corename)/schema/fields
> /solr/(corename)/schema/fields/fieldname
> /solr/(corename)/schema/dynamicfields
> /solr/(corename)/schema/dynamicfields/pattern
> /solr/(corename)/schema/fieldtypes
> /solr/(corename)/schema/fieldtypes/typename 
> /solr/(corename)/schema/copyfields

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Created] (SOLR-4526) Admin UI dependence on UnixOperationSystemMXBean

2013-03-04 Thread Shawn Heisey (JIRA)
Shawn Heisey created SOLR-4526:
--

 Summary: Admin UI dependence on UnixOperationSystemMXBean
 Key: SOLR-4526
 URL: https://issues.apache.org/jira/browse/SOLR-4526
 Project: Solr
  Issue Type: Bug
  Components: web gui
Affects Versions: 4.1
 Environment: Oracle Corporation OpenJDK 64-Bit Server VM (1.7.0_15 
23.7-b01)
Reporter: Shawn Heisey


A user on IRC was having trouble getting file descriptor counts and JVM memory 
usage in the admin UI, but it worked perfectly fine on another system.  The 
problem system uses OpenJDK, the other one uses the Apple JDK.  The user had 
tracked it down to an exception while trying to get open file descriptor info.

Looking in the SystemInfoHandler.java file, I see a comment reference to 
com.sun.management.UnixOperatingSystemMXBean at the point where it is getting 
file descriptor info.  A little extra searching turned up ZOOKEEPER-1579 which 
refers to HBASE-6495, the same problem with OpenJDK.


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Updated] (SOLR-4503) Add REST API methods to get schema information: fields, dynamic fields, and field types

2013-03-04 Thread Steve Rowe (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4503?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Rowe updated SOLR-4503:
-

Attachment: SOLR-4503.patch
all.field.types.json
all.fields.json
all.dynamic.fields.json
coordinate.dynamic.field.json
date.field.type.json
price.field.json
copyfields.json

Patch, I think it's ready to go.  I've also attached updated example outputs.

I've split off a new /schema/copyfields/ request, because it wasn't possible to 
attach copyFields with subset pattern dynamic field references off of anything 
in the /schema/fields/ structure.  /schema/copyfields/ also now contains 
maxChars, sourceDynamicBase and destDynamicBase, if applicable.  I had to 
refactor dynamic field handling in IndexSchema.java in order to fix a bug 
identified in SOLR-3798 - this refactoring+fix is included in this patch.

By default, all requests now exclude default properties.  A "showDefaults" 
query parameter causes them to be included in the response.

By default, all requests are indented JSON.

There are tests for everything, 'ant test' passes under Solr, 'ant precommit' 
passes, and I've added CHANGES.txt entries.

> Add REST API methods to get schema information: fields, dynamic fields, and 
> field types
> ---
>
> Key: SOLR-4503
> URL: https://issues.apache.org/jira/browse/SOLR-4503
> Project: Solr
>  Issue Type: Sub-task
>  Components: Schema and Analysis
>Affects Versions: 4.1
>Reporter: Steve Rowe
>Assignee: Steve Rowe
> Attachments: all.dynamic.fields.json, all.dynamic.fields.json, 
> all.fields.json, all.fields.json, all.field.types.json, all.field.types.json, 
> coordinate.dynamic.field.json, coordinate.dynamic.field.json, 
> copyfields.json, date.field.type.json, date.field.type.json, 
> price.field.json, price.field.json, SOLR-4503.patch, SOLR-4503.patch, 
> SOLR-4503.patch
>
>
> Add REST methods that provide properties for fields, dynamic fields, and 
> field types, using paths:
> /solr/(corename)/schema/fields
> /solr/(corename)/schema/fields/fieldname
> /solr/(corename)/schema/dynamicfields
> /solr/(corename)/schema/dynamicfields/pattern
> /solr/(corename)/schema/fieldtypes
> /solr/(corename)/schema/fieldtypes/typename 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4752) Merge segments to sort them

2013-03-04 Thread Adrien Grand (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4752?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592366#comment-13592366
 ] 

Adrien Grand commented on LUCENE-4752:
--

bq. How can you early terminate a query for a single segment? [...] To early 
terminate efficiently, you must have the segments in a consistent order, e.g. 
S1 > S2 > S3.

I think this is just an API limitation? Segments being processed independently, 
we should be able to terminate collection on a per-segment basis? 

bq. instead of stuffing into IWC what seems like a random setting 
(pick-segments-for-sorting), we should have something more generic, like 
AtomicReaderFactory

I didn't mean this should be a boolean. Of course it should be something more 
flexible/configurable! I'm very bad at picking names, but following your naming 
suggestion, we could have something like
{code}
abstract class AtomicReaderFactory {
  abstract List reorder(List segmentReaders);
}
{code}?

The default impl would be the identity whereas the sorting impl would return a 
singleton containing a sorted view over the segment readers?

bq. Also, a custom SegmentMerger to implement the zig-zag merge would help too.

This is another option. I actually started exploring this option when David 
opened this issue, but it can become complicated, especially for postings lists 
merging, whereas reusing the sorted view from LUCENE-3918 would make merging 
trivial.

> Merge segments to sort them
> ---
>
> Key: LUCENE-4752
> URL: https://issues.apache.org/jira/browse/LUCENE-4752
> Project: Lucene - Core
>  Issue Type: New Feature
>  Components: core/index
>Reporter: David Smiley
>Assignee: Adrien Grand
>
> It would be awesome if Lucene could write the documents out in a segment 
> based on a configurable order.  This of course applies to merging segments 
> to. The benefit is increased locality on disk of documents that are likely to 
> be accessed together.  This often applies to documents near each other in 
> time, but also spatially.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4752) Merge segments to sort them

2013-03-04 Thread David Smiley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4752?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592362#comment-13592362
 ] 

David Smiley commented on LUCENE-4752:
--

I wonder what other big-data software is also sorting its data files both 
within-file *and* across them (the latter being the tricker part I think)?  
Cassandra, HBase, or Accumulo?  The code details are going to be specific to 
the platform but I'm interested in the scheduling / merging algorithm, which 
seems like the biggest challenge to me.  I bet this has been solved before.  My 
initial attempts at coming up with an algorithm on my notepad isn't showing 
much promise.

> Merge segments to sort them
> ---
>
> Key: LUCENE-4752
> URL: https://issues.apache.org/jira/browse/LUCENE-4752
> Project: Lucene - Core
>  Issue Type: New Feature
>  Components: core/index
>Reporter: David Smiley
>Assignee: Adrien Grand
>
> It would be awesome if Lucene could write the documents out in a segment 
> based on a configurable order.  This of course applies to merging segments 
> to. The benefit is increased locality on disk of documents that are likely to 
> be accessed together.  This often applies to documents near each other in 
> time, but also spatially.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4752) Merge segments to sort them

2013-03-04 Thread Shai Erera (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4752?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592355#comment-13592355
 ] 

Shai Erera commented on LUCENE-4752:


I was commenting on Adrien's statement about how this would benefit early 
termination. If we open up the hooks (AtomicReaderFactory and custom 
SegmentMerger), I think that will open the way for many interesting things to 
do during merging, including merge-sorting segments.

> Merge segments to sort them
> ---
>
> Key: LUCENE-4752
> URL: https://issues.apache.org/jira/browse/LUCENE-4752
> Project: Lucene - Core
>  Issue Type: New Feature
>  Components: core/index
>Reporter: David Smiley
>Assignee: Adrien Grand
>
> It would be awesome if Lucene could write the documents out in a segment 
> based on a configurable order.  This of course applies to merging segments 
> to. The benefit is increased locality on disk of documents that are likely to 
> be accessed together.  This often applies to documents near each other in 
> time, but also spatially.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4752) Merge segments to sort them

2013-03-04 Thread David Smiley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4752?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592346#comment-13592346
 ] 

David Smiley commented on LUCENE-4752:
--

Shai, Early-termination is a non-goal for this issue.  Locality of reference 
is, with further index compression being a bonus.

> Merge segments to sort them
> ---
>
> Key: LUCENE-4752
> URL: https://issues.apache.org/jira/browse/LUCENE-4752
> Project: Lucene - Core
>  Issue Type: New Feature
>  Components: core/index
>Reporter: David Smiley
>Assignee: Adrien Grand
>
> It would be awesome if Lucene could write the documents out in a segment 
> based on a configurable order.  This of course applies to merging segments 
> to. The benefit is increased locality on disk of documents that are likely to 
> be accessed together.  This often applies to documents near each other in 
> time, but also spatially.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



SolrJ client changes

2013-03-04 Thread Kirk True
Hi all,

First off, thanks for the great project. We're really excited about
incorporating Solr into our stack. Thanks!

We're looking at making some changes to the SolrJ client to meet our
operational needs:


   - routing of requests to specific nodes for better cache hit ratios
   - adding instrumentation via JMX (list of availability of Solr servers,
   response time, various counters)
- implementing connection pooling (failure detection and recovery);
   this is probably achievable on a layer atop the SolrJ client vs. as a part
   of it
- logging errors more consistently
   - miscellaneous changes (debugging, use of POST, optional custom
   User-Agent, etc.)


So far, it appears that these changes can be made by adding hooks or
optional APIs to SolrJ that we would then implement with our custom logic.
So there's no secret sauce or anything.

We are curious for your recommendations about how best to make these
changes.

A few questions:


   1. It seems that the SolrJ client is part of the overall Solr project,
   so it would be harder/slower for us to get our changes in, committed, and
   released in a timely manner.
   2. We're specifically interested in the load balancing SolrJ client
   (LBHttpSolrServer) but it doesn't appear to be recommended for production
   use. Thoughts on its usage?
   3. Are there perhaps other clients for Solr for Java other than SolrJ
   and scalikesolr?


Thanks,
Kirk


[jira] [Commented] (LUCENE-3918) Port index sorter to trunk APIs

2013-03-04 Thread David Smiley (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3918?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592301#comment-13592301
 ] 

David Smiley commented on LUCENE-3918:
--

bq. Global ordering over a multi-segment index seems less useful than the local 
per-segment ordering - after all, the main use cases are compression 
improvements and early termination strategies, both of which operate on a 
per-segment basis.

Those are two "main use-cases" but there is a third that (to me) is far more 
important than either -- locality of reference.  Retrieving documents near each 
other (according to the configured sort order) should ideally have their stored 
fields data very close together.  The early-termination use-case is clearly 
important to you but is not the goal of LUCENE-4752.

> Port index sorter to trunk APIs
> ---
>
> Key: LUCENE-3918
> URL: https://issues.apache.org/jira/browse/LUCENE-3918
> Project: Lucene - Core
>  Issue Type: Task
>  Components: modules/other
>Affects Versions: 4.0-ALPHA
>Reporter: Robert Muir
> Fix For: 4.2, 5.0
>
> Attachments: LUCENE-3918.patch, LUCENE-3918.patch, LUCENE-3918.patch, 
> LUCENE-3918.patch, LUCENE-3918.patch, LUCENE-3918.patch, LUCENE-3918.patch, 
> LUCENE-3918.patch
>
>
> LUCENE-2482 added an IndexSorter to 3.x, but we need to port this
> functionality to 4.0 apis.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4752) Merge segments to sort them

2013-03-04 Thread Shai Erera (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4752?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592289#comment-13592289
 ] 

Shai Erera commented on LUCENE-4752:


How can you early terminate a query for a single segment? Say that you have 3 
sorted segments (individually) and your query asks to get the top-10 of some 
criteria. The top-10 may come from the 3 segments as follows: seg1=4, seg2=4, 
seg3=2. But you don't know that until you processed all 3 segments right? While 
you could make a decision on a per-segment basis to 'terminate', there's no 
mechanism today to tell IndexSearcher "I'm done w/ that segment, move on". 
Today, if you want to early terminate, you need to throw an exception from the 
Collector, and catch it outside, in your application code?

To early terminate efficiently, you must have the segments in a consistent 
order, e.g. S1 > S2 > S3. Then, after you've processed enough elements from S1, 
you can early terminate the entire query because you're guaranteed that 
successive documents will be "smaller".

Unless, you add to your Collector.collect() an "if (done) return" and consider 
that a no-op, or make your own IndexSearcher logic ... then per-segment early 
termination is doable.

As for the approach you describe, I think that instead of stuffing into IWC 
what seems like a random setting (pick-segments-for-sorting), we should have 
something more generic, like AtomicReaderFactory, which IW will use instead of 
always loading SegmentReader. That will let you load your custom AtomicReader? 
Or, perhaps this can be a SortingCodec? Also, a custom SegmentMerger to 
implement the zig-zag merge would help too.

> Merge segments to sort them
> ---
>
> Key: LUCENE-4752
> URL: https://issues.apache.org/jira/browse/LUCENE-4752
> Project: Lucene - Core
>  Issue Type: New Feature
>  Components: core/index
>Reporter: David Smiley
>Assignee: Adrien Grand
>
> It would be awesome if Lucene could write the documents out in a segment 
> based on a configurable order.  This of course applies to merging segments 
> to. The benefit is increased locality on disk of documents that are likely to 
> be accessed together.  This often applies to documents near each other in 
> time, but also spatially.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: verbose logging

2013-03-04 Thread Dawid Weiss
By default the output from tests that passed is not dumped to the
console (you wouldn't be able to track which JVM dumped what anyway
when running on multicore). If you need a full log then:

> ant test-help
...
 [echo] # Include additional information like what is printed to
 [echo] # sysout/syserr, even if the test passes.
 [echo] # Enabled automatically when running for a single test case.
 [echo] ant -Dtests.showSuccess=true test

Should do the job. If you'd like to see the output from jenkins or any
other build server then, unless it fails, it won't be on the console.

Dawid

On Mon, Mar 4, 2013 at 3:48 PM, Erick Erickson  wrote:
> I tried turning on verbose in the OpenCloseCoreStressTests by setting
> System.setProperty("tests.verbose", "true");
> in @BeforeClass and, of course, it works on my machine. But the failed test
> logs don't have verbose output, is there some magic here? For the nonce, I'm
> just doing a bit of hard logging.
>
> I also found another place where I screwed up the a Stream by not closing
> it, ZkCLI of all places. I've checked that in.
>
> Of course all tests pass on my machine. Scant comfort that.
>
> I also had yet another assumption proved wrong so I didn't get any useful
> info from last night's failure, trying to add some more logging again today.
> Don't panic that the checking has my initials init, that's so I'm absolutely
> sure to know to take it out.
>
> Thanks,
> Erick

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4505) Deadlock around SolrCoreState update lock.

2013-03-04 Thread Erick Erickson (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4505?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592250#comment-13592250
 ] 

Erick Erickson commented on SOLR-4505:
--

Tagged http://svn.apache.org/viewvc?view=revision&revision=1452321 to the wrong 
JIRA, should have been SOLR-4525

> Deadlock around SolrCoreState update lock.
> --
>
> Key: SOLR-4505
> URL: https://issues.apache.org/jira/browse/SOLR-4505
> Project: Solr
>  Issue Type: Bug
>Reporter: Mark Miller
>Assignee: Mark Miller
>Priority: Minor
> Fix For: 4.2, 5.0
>
> Attachments: newstack.txt, newstack.txt, newstack.txt, 
> SOLR-4505.patch, SOLR-4505.patch, SOLR-4505.patch, SOLR-4505.patch
>
>
> Erick found a deadlock with his core stress tool - see 
> http://markmail.org/message/aq5hghbqia2uimgl

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



verbose logging

2013-03-04 Thread Erick Erickson
I tried turning on verbose in the OpenCloseCoreStressTests by setting
System.setProperty("tests.verbose", "true");
in @BeforeClass and, of course, it works on my machine. But the failed test
logs don't have verbose output, is there some magic here? For the nonce,
I'm just doing a bit of hard logging.

I also found another place where I screwed up the a Stream by not closing
it, ZkCLI of all places. I've checked that in.

Of course all tests pass on my machine. Scant comfort that.

I also had yet another assumption proved wrong so I didn't get any useful
info from last night's failure, trying to add some more logging again
today. Don't panic that the checking has my initials init, that's so I'm
absolutely sure to know to take it out.

Thanks,
Erick


[jira] [Commented] (LUCENE-3918) Port index sorter to trunk APIs

2013-03-04 Thread Shai Erera (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3918?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592237#comment-13592237
 ] 

Shai Erera commented on LUCENE-3918:


I agree that per-segment sorting may improve compression, but I don't see how 
it can help with early termination. If Lucene allowed you to terminate on a 
per-segment basis then maybe, but today (as far as I know), the only way to 
early terminate a query is by throwing an exception and catching it outside, 
right?

At any rate, per-segment sorting doesn't help much when segments are merged, 
because then you lose sorting again. Unless, we add some API to IW / 
SegmentMerger for you to control how the segments are merged...

> Port index sorter to trunk APIs
> ---
>
> Key: LUCENE-3918
> URL: https://issues.apache.org/jira/browse/LUCENE-3918
> Project: Lucene - Core
>  Issue Type: Task
>  Components: modules/other
>Affects Versions: 4.0-ALPHA
>Reporter: Robert Muir
> Fix For: 4.2, 5.0
>
> Attachments: LUCENE-3918.patch, LUCENE-3918.patch, LUCENE-3918.patch, 
> LUCENE-3918.patch, LUCENE-3918.patch, LUCENE-3918.patch, LUCENE-3918.patch, 
> LUCENE-3918.patch
>
>
> LUCENE-2482 added an IndexSorter to 3.x, but we need to port this
> functionality to 4.0 apis.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4505) Deadlock around SolrCoreState update lock.

2013-03-04 Thread Commit Tag Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4505?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592229#comment-13592229
 ] 

Commit Tag Bot commented on SOLR-4505:
--

[trunk commit] Erick Erickson
http://svn.apache.org/viewvc?view=revision&revision=1452321

SOLR-4505, another place where I missed closing the stream


> Deadlock around SolrCoreState update lock.
> --
>
> Key: SOLR-4505
> URL: https://issues.apache.org/jira/browse/SOLR-4505
> Project: Solr
>  Issue Type: Bug
>Reporter: Mark Miller
>Assignee: Mark Miller
>Priority: Minor
> Fix For: 4.2, 5.0
>
> Attachments: newstack.txt, newstack.txt, newstack.txt, 
> SOLR-4505.patch, SOLR-4505.patch, SOLR-4505.patch, SOLR-4505.patch
>
>
> Erick found a deadlock with his core stress tool - see 
> http://markmail.org/message/aq5hghbqia2uimgl

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



RE: [JENKINS] Lucene-Solr-4.x-Linux (32bit/jdk1.8.0-ea-b79) - Build # 4533 - Still Failing!

2013-03-04 Thread Uwe Schindler
> > The parent process or the worker child?
> 
> The parent. The reason here was to avoid spawning a background thread on
> the child to avoid any interference with existing tests. So the parent tracks
> events that tell where the child is going (entering or leaving a suite/ test) 
> and
> is from this deducing where a child may be.
> The  status means the child is not executing any suite; it used to
> be the case that OOM-ed code would hang forever on System.exit, for
> example...
> 
> > Can you see from the messages *where* it might hang?
> 
> Not really, not if it's a JVM issue.
> 
> > Interestingly another one is now hanging with the following message:
> >
> > [junit4:junit4] HEARTBEAT J0 PID(11405@serv1): 2013-03-04T13:16:16,
> > stalled for 7086s at:
> > UIMABaseAnalyzerTest.baseUIMAAnalyzerIntegrationTest
> 
> If this one is not responsive to regular stack trace probing then it's 
> definitely a
> JVM freeze somewhere.

It was also frozen! Was only possible to kill -9. Unfortunately attaching with 
gdb did not work to me, I'll try next time again.

Uwe


-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



Re: [JENKINS] Lucene-Solr-4.x-Linux (32bit/jdk1.8.0-ea-b79) - Build # 4533 - Still Failing!

2013-03-04 Thread Dawid Weiss
> The parent process or the worker child?

The parent. The reason here was to avoid spawning a background thread
on the child to avoid any interference with existing tests. So the
parent tracks events that tell where the child is going (entering or
leaving a suite/ test) and is from this deducing where a child may be.
The  status means the child is not executing any suite; it
used to be the case that OOM-ed code would hang forever on
System.exit, for example...

> Can you see from the messages *where* it might hang?

Not really, not if it's a JVM issue.

> Interestingly another one is now hanging with the following message:
>
> [junit4:junit4] HEARTBEAT J0 PID(11405@serv1): 2013-03-04T13:16:16, stalled 
> for 7086s at: UIMABaseAnalyzerTest.baseUIMAAnalyzerIntegrationTest

If this one is not responsive to regular stack trace probing then it's
definitely a JVM freeze somewhere.

D.

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-4.x-Linux (32bit/jdk1.8.0-ea-b79) - Build # 4535 - Failure!

2013-03-04 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/4535/
Java: 32bit/jdk1.8.0-ea-b79 -server -XX:+UseG1GC

All tests passed

Build Log:
[...truncated 6055 lines...]
[junit4:junit4] ERROR: JVM J0 ended with an exception, command line: 
/var/lib/jenkins/tools/java/32bit/jdk1.8.0-ea-b79/jre/bin/java -server 
-XX:+UseG1GC -XX:+HeapDumpOnOutOfMemoryError 
-XX:HeapDumpPath=/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/heapdumps 
-Dtests.prefix=tests -Dtests.seed=D1F2BAB08DD89503 -Xmx512M -Dtests.iters= 
-Dtests.verbose=false -Dtests.infostream=false -Dtests.codec=random 
-Dtests.postingsformat=random -Dtests.docvaluesformat=random 
-Dtests.locale=random -Dtests.timezone=random -Dtests.directory=random 
-Dtests.linedocsfile=europarl.lines.txt.gz -Dtests.luceneMatchVersion=4.2 
-Dtests.cleanthreads=perMethod 
-Djava.util.logging.config.file=/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/tools/junit4/logging.properties
 -Dtests.nightly=false -Dtests.weekly=false -Dtests.slow=true 
-Dtests.asserts.gracious=false -Dtests.multiplier=3 -DtempDir=. 
-Djava.io.tmpdir=. 
-Djunit4.tempDir=/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build/analysis/uima/test/temp
 
-Dclover.db.dir=/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build/clover/db
 -Djava.security.manager=org.apache.lucene.util.TestSecurityManager 
-Djava.security.policy=/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/tools/junit4/tests.policy
 -Dlucene.version=4.2-SNAPSHOT -Djetty.testMode=1 -Djetty.insecurerandom=1 
-Dsolr.directoryFactory=org.apache.solr.core.MockDirectoryFactory 
-Djava.awt.headless=true -Dfile.encoding=ISO-8859-1 -classpath 
/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build/analysis/uima/classes/test:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build/test-framework/classes/java:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build/codecs/classes/java:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build/analysis/common/lucene-analyzers-common-4.2-SNAPSHOT.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/analysis/uima/src/test-files:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/analysis/uima/lib/Tagger-2.3.1.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/analysis/uima/lib/WhitespaceTokenizer-2.3.1.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/analysis/uima/lib/uimaj-core-2.3.1.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build/core/classes/java:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/test-framework/lib/junit-4.10.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/test-framework/lib/randomizedtesting-runner-2.0.8.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/build/analysis/uima/classes/java:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-launcher.jar:/var/lib/jenkins/.ant/lib/ivy-2.3.0.jar:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-jai.jar:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-swing.jar:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-apache-oro.jar:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-jmf.jar:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-apache-xalan2.jar:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-javamail.jar:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-apache-resolver.jar:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-testutil.jar:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-commons-logging.jar:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-apache-log4j.jar:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-junit.jar:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-jsch.jar:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-commons-net.jar:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-apache-bsf.jar:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-jdepend.jar:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-netrexx.jar:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-apache-regexp.jar:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-junit4.jar:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant.jar:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-apache-bcel.jar:/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-antlr.jar:/var/lib/jenkins/tools/java/32bit/jdk1.8.0-ea-b79/lib/tools.jar:/var/lib/jenkins/.ivy2/cache/com.carrotsearch.randomizedtesting/junit4-ant/jars/junit4-ant-2.0.8.jar
 -ea:org.apache.lucene... -ea:org.apache.solr... 
com.carrotsearch.ant.tasks.junit4.sla

RE: [JENKINS] Lucene-Solr-4.x-Linux (32bit/jdk1.8.0-ea-b79) - Build # 4533 - Still Failing!

2013-03-04 Thread Uwe Schindler
I can try to hook it up with gdb and report to the hotspot mailing list. 
Unfortunately, we have nothing to report, only that the JVM hangs. Who is 
printing the messages? The parent process or the worker child? If the parent 
prints them, the child can be dead altogether, otherwise it might still be able 
to print messages. Can you see from the messages *where* it might hang?

[junit4:junit4] HEARTBEAT J0 PID(6790@serv1): 2013-03-04T08:01:06, stalled for 
2464s at: 

Interestingly another one is now hanging with the following message:

[junit4:junit4] HEARTBEAT J0 PID(11405@serv1): 2013-03-04T13:16:16, stalled for 
7086s at: UIMABaseAnalyzerTest.baseUIMAAnalyzerIntegrationTest

Uwe

-
Uwe Schindler
H.-H.-Meier-Allee 63, D-28213 Bremen
http://www.thetaphi.de
eMail: u...@thetaphi.de


> -Original Message-
> From: dawid.we...@gmail.com [mailto:dawid.we...@gmail.com] On Behalf
> Of Dawid Weiss
> Sent: Monday, March 04, 2013 9:12 AM
> To: dev@lucene.apache.org
> Subject: Re: [JENKINS] Lucene-Solr-4.x-Linux (32bit/jdk1.8.0-ea-b79) - Build #
> 4533 - Still Failing!
> 
> It'd be nice to get a native stack/ core dump of this via gdb and report back 
> to
> the hotspot folks. I know too little about gdb to help here but I bet it's
> possible.
> 
> Dawid
> 
> On Mon, Mar 4, 2013 at 9:03 AM, Uwe Schindler  wrote:
> > Killed again! Again 32 bit JDK8 with G1 Collector
> >
> > -
> > Uwe Schindler
> > H.-H.-Meier-Allee 63, D-28213 Bremen
> > http://www.thetaphi.de
> > eMail: u...@thetaphi.de
> >
> >> -Original Message-
> >> From: Policeman Jenkins Server [mailto:jenk...@thetaphi.de]
> >> Sent: Monday, March 04, 2013 9:02 AM
> >> To: dev@lucene.apache.org
> >> Subject: [JENKINS] Lucene-Solr-4.x-Linux (32bit/jdk1.8.0-ea-b79) -
> >> Build #
> >> 4533 - Still Failing!
> >>
> >> Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Linux/4533/
> >> Java: 32bit/jdk1.8.0-ea-b79 -client -XX:+UseG1GC
> >>
> >> All tests passed
> >>
> >> Build Log:
> >> [...truncated 5917 lines...]
> >> [junit4:junit4] ERROR: JVM J0 ended with an exception, command line:
> >> /var/lib/jenkins/tools/java/32bit/jdk1.8.0-ea-b79/jre/bin/java
> >> -client - XX:+UseG1GC -XX:+HeapDumpOnOutOfMemoryError -
> >> XX:HeapDumpPath=/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-
> >> Linux/heapdumps -Dtests.prefix=tests -Dtests.seed=3B99A947D2A6DD6D
> -
> >> Xmx512M -Dtests.iters= -Dtests.verbose=false -Dtests.infostream=false
> >> - Dtests.codec=random -Dtests.postingsformat=random -
> >> Dtests.docvaluesformat=random -Dtests.locale=random -
> >> Dtests.timezone=random -Dtests.directory=random -
> >> Dtests.linedocsfile=europarl.lines.txt.gz
> >> -Dtests.luceneMatchVersion=4.2 - Dtests.cleanthreads=perMethod -
> >> Djava.util.logging.config.file=/mnt/ssd/jenkins/workspace/Lucene-Solr
> >> -4.x- Linux/lucene/tools/junit4/logging.properties
> >> -Dtests.nightly=false - Dtests.weekly=false -Dtests.slow=true
> >> -Dtests.asserts.gracious=false -
> >> Dtests.multiplier=3 -DtempDir=. -Djava.io.tmpdir=. -
> >> Djunit4.tempDir=/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-
> >> Linux/lucene/build/analysis/uima/test/temp -
> >> Dclover.db.dir=/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-
> >> Linux/lucene/build/clover/db -
> >> Djava.security.manager=org.apache.lucene.util.TestSecurityManager -
> >> Djava.security.policy=/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-
> >> Linux/lucene/tools/junit4/tests.policy -Dlucene.version=4.2-SNAPSHOT
> >> -
> >> Djetty.testMode=1 -Djetty.insecurerandom=1 -
> >> Dsolr.directoryFactory=org.apache.solr.core.MockDirectoryFactory -
> >> Djava.awt.headless=true -classpath
> /mnt/ssd/jenkins/workspace/Lucene-
> >> Solr-4.x-
> >> Linux/lucene/build/analysis/uima/classes/test:/mnt/ssd/jenkins/worksp
> >> ace
> >> /Lucene-Solr-4.x-Linux/lucene/build/test-
> >> framework/classes/java:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-
> >> Linux/lucene/build/codecs/classes/java:/mnt/ssd/jenkins/workspace/Luc
> >> en
> >> e-Solr-4.x-Linux/lucene/build/analysis/common/lucene-analyzers-
> common
> >> -
> >> 4.2-SNAPSHOT.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-
> >> Linux/lucene/analysis/uima/src/test-
> >> files:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-
> >> Linux/lucene/analysis/uima/lib/Tagger-
> >> 2.3.1.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-
> >> Linux/lucene/analysis/uima/lib/WhitespaceTokenizer-
> >> 2.3.1.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-
> >> Linux/lucene/analysis/uima/lib/uimaj-core-
> >> 2.3.1.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-
> >>
> Linux/lucene/build/core/classes/java:/mnt/ssd/jenkins/workspace/Lucen
> >> e-
> >> Solr-4.x-Linux/lucene/test-framework/lib/junit-
> >> 4.10.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-Linux/lucene/test
> >> -
> >> framework/lib/randomizedtesting-runner-
> >> 2.0.8.jar:/mnt/ssd/jenkins/workspace/Lucene-Solr-4.x-
> >> Linux/lucene/build/analysis/uima/classes/java:/var/lib/jenkins/tools/
> >> hudson
> >> .tasks.Ant_AntInstallation/ANT_1.8.2/lib/ant-
> >>

[jira] [Updated] (LUCENE-3918) Port index sorter to trunk APIs

2013-03-04 Thread Shai Erera (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-3918?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shai Erera updated LUCENE-3918:
---

Attachment: LUCENE-3918.patch

Patch:

* Consolidates tests into less classes, also creates the test index in 
BeforeClass

* Randomizing tests: #docs, sort permutations

* Added _TestUtil.checkIndex and checkReader (see below) - this uncovered few 
bugs

* Moved relevant classes to inherit from their FilterXYZ counterpart - reduced 
code

In SortingAtomicReaderTest I had to call _TestUtil.checkReader with 
crossCheckTermVectors=false. I put a comment why, but I'll repeat it here. 
Let's say that your permutation is [2,0,1] which means doc 0 goes to 2, 1 goes 
to 0 and 2 goes to 1. You then ask for termVectors(0), and receive, correctly, 
those of document 2. Now, when crossCheck validation occurs, it iterates on the 
terms of the TV, expecting to find each term associated with document 0 (since 
CheckIndex doesn't know about sorting reader, and that in fact it works now w/ 
doc #2). However, the terms of doc #2 are mapped, again - correctly, to 
document #1 (as the permutation specifies). Therefore CheckIndex fails since it 
expects doc 0 but receives doc 1.

Unless I misunderstand how this cross-checking should work, I think that it's 
ok to disable cross-checking on the reader, in this case. In SorterUtilTest, 
where I addIndexes, I do checkIndex with cross-checking, and there it passes 
(as it should).

All that's missing is CHANGES, but I'll let others review ...

> Port index sorter to trunk APIs
> ---
>
> Key: LUCENE-3918
> URL: https://issues.apache.org/jira/browse/LUCENE-3918
> Project: Lucene - Core
>  Issue Type: Task
>  Components: modules/other
>Affects Versions: 4.0-ALPHA
>Reporter: Robert Muir
> Fix For: 4.2, 5.0
>
> Attachments: LUCENE-3918.patch, LUCENE-3918.patch, LUCENE-3918.patch, 
> LUCENE-3918.patch, LUCENE-3918.patch, LUCENE-3918.patch, LUCENE-3918.patch, 
> LUCENE-3918.patch
>
>
> LUCENE-2482 added an IndexSorter to 3.x, but we need to port this
> functionality to 4.0 apis.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-4654) Test duration statistics from multiple test runs should be reused (locally).

2013-03-04 Thread Commit Tag Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4654?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592172#comment-13592172
 ] 

Commit Tag Bot commented on LUCENE-4654:


[branch_4x commit] Dawid Weiss
http://svn.apache.org/viewvc?view=revision&revision=1452277

LUCENE-4654: Test duration statistics from multiple test runs should be reused 
(locally).


> Test duration statistics from multiple test runs should be reused (locally).
> 
>
> Key: LUCENE-4654
> URL: https://issues.apache.org/jira/browse/LUCENE-4654
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Trivial
> Fix For: 4.2, 5.0
>
>
> This is trivial to accomplish: when somebody (or jenkins) runs tests multiple 
> times the execution statistics could be reused to improve load balancing on 
> the local machine (local hardware and settings) in favor of the precached 
> values currently version in the svn repo.
> At this moment we already do this, but keep the stats under build/ and every 
> ant clean effectively removes them. I could move those stats under an 
> svn-ignored folder elsewhere so that these stats are not lost and reused for 
> balancing.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-NightlyTests-trunk - Build # 196 - Still Failing

2013-03-04 Thread Apache Jenkins Server
Build: https://builds.apache.org/job/Lucene-Solr-NightlyTests-trunk/196/

All tests passed

Build Log:
[...truncated 29229 lines...]
BUILD FAILED
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-NightlyTests-trunk/build.xml:388:
 The following error occurred while executing this line:
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-NightlyTests-trunk/build.xml:320:
 The following error occurred while executing this line:
/usr/home/hudson/hudson-slave/workspace/Lucene-Solr-NightlyTests-trunk/extra-targets.xml:120:
 The following files are missing svn:eol-style (or binary svn:mime-type):
* solr/licenses/carrot2-mini-LICENSE-BSD_LIKE.txt
* solr/licenses/carrot2-mini-NOTICE.txt

Total time: 107 minutes 14 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Recording test results
Email was triggered for: Failure
Sending email for trigger: Failure



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Commented] (LUCENE-4654) Test duration statistics from multiple test runs should be reused (locally).

2013-03-04 Thread Commit Tag Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4654?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592165#comment-13592165
 ] 

Commit Tag Bot commented on LUCENE-4654:


[trunk commit] Dawid Weiss
http://svn.apache.org/viewvc?view=revision&revision=1452276

LUCENE-4654: Test duration statistics from multiple test runs should be reused 
(locally).


> Test duration statistics from multiple test runs should be reused (locally).
> 
>
> Key: LUCENE-4654
> URL: https://issues.apache.org/jira/browse/LUCENE-4654
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Trivial
> Fix For: 4.2, 5.0
>
>
> This is trivial to accomplish: when somebody (or jenkins) runs tests multiple 
> times the execution statistics could be reused to improve load balancing on 
> the local machine (local hardware and settings) in favor of the precached 
> values currently version in the svn repo.
> At this moment we already do this, but keep the stats under build/ and every 
> ant clean effectively removes them. I could move those stats under an 
> svn-ignored folder elsewhere so that these stats are not lost and reused for 
> balancing.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-4654) Test duration statistics from multiple test runs should be reused (locally).

2013-03-04 Thread Dawid Weiss (JIRA)

 [ 
https://issues.apache.org/jira/browse/LUCENE-4654?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dawid Weiss resolved LUCENE-4654.
-

Resolution: Fixed

I've implemented this and committed it in along with an update to precomputed 
statistics.

This will *not* make your tests a whole lot faster because there is 
job-stealing by default anyway so unless you really hit a long test scheduled 
at the very end of a series of per-jvm tests it won't make a difference.

Obviously it won't make any difference if you're running with a single JVM 
either.

What may make a (small) difference is for people with a lot of cores or, 
paradoxically, a few cores on slower machines. I think it's an improvement over 
the existing version anyway so it's for the better.


> Test duration statistics from multiple test runs should be reused (locally).
> 
>
> Key: LUCENE-4654
> URL: https://issues.apache.org/jira/browse/LUCENE-4654
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Trivial
> Fix For: 4.2, 5.0
>
>
> This is trivial to accomplish: when somebody (or jenkins) runs tests multiple 
> times the execution statistics could be reused to improve load balancing on 
> the local machine (local hardware and settings) in favor of the precached 
> values currently version in the svn repo.
> At this moment we already do this, but keep the stats under build/ and every 
> ant clean effectively removes them. I could move those stats under an 
> svn-ignored folder elsewhere so that these stats are not lost and reused for 
> balancing.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4503) Add REST API methods to get schema information: fields, dynamic fields, and field types

2013-03-04 Thread Steve Rowe (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4503?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592122#comment-13592122
 ] 

Steve Rowe commented on SOLR-4503:
--

Thanks for the review, Yonik.

bq. It seems like there are multiple use-cases for the fields API...

{quote}
One just wants to get all the info/properties for a field, and they don't care 
how that field is defined.
For example, they may want to look up field X to see if it's indexed, w/o 
having to worry about if it's a "dynamic" field or not.
{quote}

I originally had code to do index lookup, but this API is about the schema, so 
I took it out, thinking that index queries didn't belong.

bq. allowing the fields API to be used for dynamic fields also, but provide an 
indicator about what pattern matched

In my current state I do have an "includeDynamic" query param for the 
/schema/fields/fieldname request, to report the matching dynamic field 
properties if fieldname isn't an explicitly declared field.

{quote}
One wants to get the actual definition definition in the schema (or as close to 
that as possible), and they would expect to see something very close to what 
they defined (or sent in using the fields API to create a new field). Showing 
all possible field properties like "storeOffsetsWithPositions" are going to be 
potentially confusing since people won't know what they mean and when/why they 
have to specify them when creating new fields.
{quote}

bq. only returning non-default properties... so if you create an integer field 
with "indexed=true", that's pretty much all you get back.
bq. having a parameter to allow requesting all "look through" properties.

Good idea, I'll do this.

{quote}
Also related: SOLR-4210
We should aim for being able to hit any node in the cluster w/o worrying about 
which nodes are hosting which collections.
{quote}

Skimming Mark's patch modifying SolrDispatchFilter, I think the schema REST 
requests will already be proxied, but I'll test to be sure.

bq. We should probably also default to indented JSON output.

OK, I'll switch to that.


> Add REST API methods to get schema information: fields, dynamic fields, and 
> field types
> ---
>
> Key: SOLR-4503
> URL: https://issues.apache.org/jira/browse/SOLR-4503
> Project: Solr
>  Issue Type: Sub-task
>  Components: Schema and Analysis
>Affects Versions: 4.1
>Reporter: Steve Rowe
>Assignee: Steve Rowe
> Attachments: all.dynamic.fields.json, all.fields.json, 
> all.field.types.json, coordinate.dynamic.field.json, date.field.type.json, 
> price.field.json, SOLR-4503.patch, SOLR-4503.patch
>
>
> Add REST methods that provide properties for fields, dynamic fields, and 
> field types, using paths:
> /solr/(corename)/schema/fields
> /solr/(corename)/schema/fields/fieldname
> /solr/(corename)/schema/dynamicfields
> /solr/(corename)/schema/dynamicfields/pattern
> /solr/(corename)/schema/fieldtypes
> /solr/(corename)/schema/fieldtypes/typename 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[JENKINS] Lucene-Solr-4.x-Windows (32bit/jdk1.7.0_15) - Build # 2601 - Failure!

2013-03-04 Thread Policeman Jenkins Server
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-4.x-Windows/2601/
Java: 32bit/jdk1.7.0_15 -client -XX:+UseConcMarkSweepGC

All tests passed

Build Log:
[...truncated 30094 lines...]
BUILD FAILED
C:\Users\JenkinsSlave\workspace\Lucene-Solr-4.x-Windows\build.xml:381: The 
following error occurred while executing this line:
C:\Users\JenkinsSlave\workspace\Lucene-Solr-4.x-Windows\build.xml:320: The 
following error occurred while executing this line:
C:\Users\JenkinsSlave\workspace\Lucene-Solr-4.x-Windows\extra-targets.xml:120: 
The following files are missing svn:eol-style (or binary svn:mime-type):
* solr/licenses/carrot2-mini-LICENSE-BSD_LIKE.txt
* solr/licenses/carrot2-mini-NOTICE.txt

Total time: 72 minutes 28 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Recording test results
Description set: Java: 32bit/jdk1.7.0_15 -client -XX:+UseConcMarkSweepGC
Email was triggered for: Failure
Sending email for trigger: Failure



-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

[jira] [Commented] (LUCENE-4752) Merge segments to sort them

2013-03-04 Thread Adrien Grand (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-4752?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592107#comment-13592107
 ] 

Adrien Grand commented on LUCENE-4752:
--

I think a very simple first step could be have an experimental 
IndexWriterConfig option to tell IndexWriter to provide an atomic sorted view 
(easy once LUCENE-3918 is committed) of the segments to merge to SegmentMerger 
instead of the segments themselves (see calls to 
SegmentMerger.add(SegmentReader) in IndexWriter.mergeMiddle). Initial segments 
would remain unsorted, but the big ones, those that are interesting for both 
index compression and early query termination, would be sorted.

It can seem inefficient to sort segments over and over but I don't think we 
should worry too much:
 - if we are merging "initial" segments (those created from IndexWriter.flush), 
they would be small so sorting/merging them would be fast?
 - if we are merging big segments, I think that the following reasons could 
make merging slower than a regular merge:
   1. computing the new doc ID mapping,
   2. random I/O access,
   3. not being able to use the specialized codec merging methods.

But if the segments to merge are sorted, computing the new doc ID mapping could 
be actually fast (some sorting algorithms such as 
[TimSort|http://en.wikipedia.org/wiki/Timsort] perform in O(n) when the input 
is a succession of sorted sequences), and the access patterns to the individual 
segments would be I/O cache-friendly (because each segment would be read 
sequentially). So I think this approach could be fast enough?

> Merge segments to sort them
> ---
>
> Key: LUCENE-4752
> URL: https://issues.apache.org/jira/browse/LUCENE-4752
> Project: Lucene - Core
>  Issue Type: New Feature
>  Components: core/index
>Reporter: David Smiley
>Assignee: Adrien Grand
>
> It would be awesome if Lucene could write the documents out in a segment 
> based on a configurable order.  This of course applies to merging segments 
> to. The benefit is increased locality on disk of documents that are likely to 
> be accessed together.  This often applies to documents near each other in 
> time, but also spatially.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4470) Support for basic http auth in internal solr requests

2013-03-04 Thread Per Steffensen (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4470?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592104#comment-13592104
 ] 

Per Steffensen commented on SOLR-4470:
--

My time was prioritized to work on something else. But now I am back on this 
issue, and hopefully I will provide patch soon.

> Support for basic http auth in internal solr requests
> -
>
> Key: SOLR-4470
> URL: https://issues.apache.org/jira/browse/SOLR-4470
> Project: Solr
>  Issue Type: Bug
>  Components: clients - java, multicore, replication (java), SolrCloud
>Affects Versions: 4.0
>Reporter: Per Steffensen
>  Labels: authentication, solrclient, solrcloud
> Fix For: 4.2
>
>
> We want to protect any HTTP-resource (url). We want to require credentials no 
> matter what kind of HTTP-request you make to a Solr-node.
> It can faily easy be acheived as described on 
> http://wiki.apache.org/solr/SolrSecurity. This problem is that Solr-nodes 
> also make "internal" request to other Solr-nodes, and for it to work 
> credentials need to be provided here also.
> Ideally we would like to "forward" credentials from a particular request to 
> all the "internal" sub-requests it triggers. E.g. for search and update 
> request.
> But there are also "internal" requests
> * that only indirectly/asynchronously triggered from "outside" requests (e.g. 
> shard creation/deletion/etc based on calls to the "Collection API")
> * that do not in any way have relation to an "outside" "super"-request (e.g. 
> replica synching stuff)
> We would like to aim at a solution where "original" credentials are 
> "forwarded" when a request directly/synchronously trigger a subrequest, and 
> fallback to a configured "internal credentials" for the 
> asynchronous/non-rooted requests.
> In our solution we would aim at only supporting basic http auth, but we would 
> like to make a "framework" around it, so that not to much refactoring is 
> needed if you later want to make support for other kinds of auth (e.g. digest)
> We will work at a solution but create this JIRA issue early in order to get 
> input/comments from the community as early as possible.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (LUCENE-3918) Port index sorter to trunk APIs

2013-03-04 Thread Andrzej Bialecki (JIRA)

[ 
https://issues.apache.org/jira/browse/LUCENE-3918?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592093#comment-13592093
 ] 

Andrzej Bialecki  commented on LUCENE-3918:
---

Shai,

bq. That way, one can use it to achieve per-segment sorting .. if it means 
something for anyone (perhaps for better compression).

Global ordering over a multi-segment index seems less useful than the local 
per-segment ordering - after all, the main use cases are compression 
improvements and early termination strategies, both of which operate on a 
per-segment basis.

> Port index sorter to trunk APIs
> ---
>
> Key: LUCENE-3918
> URL: https://issues.apache.org/jira/browse/LUCENE-3918
> Project: Lucene - Core
>  Issue Type: Task
>  Components: modules/other
>Affects Versions: 4.0-ALPHA
>Reporter: Robert Muir
> Fix For: 4.2, 5.0
>
> Attachments: LUCENE-3918.patch, LUCENE-3918.patch, LUCENE-3918.patch, 
> LUCENE-3918.patch, LUCENE-3918.patch, LUCENE-3918.patch, LUCENE-3918.patch
>
>
> LUCENE-2482 added an IndexSorter to 3.x, but we need to port this
> functionality to 4.0 apis.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4259) Carrot2 dependency should be declared on the mini version, not the core

2013-03-04 Thread Commit Tag Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4259?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592087#comment-13592087
 ] 

Commit Tag Bot commented on SOLR-4259:
--

[branch_4x commit] Dawid Weiss
http://svn.apache.org/viewvc?view=revision&revision=1452221

SOLR-4259: Carrot2 dependency should be declared on the mini version, not the 
core.


> Carrot2 dependency should be declared on the mini version, not the core
> ---
>
> Key: SOLR-4259
> URL: https://issues.apache.org/jira/browse/SOLR-4259
> Project: Solr
>  Issue Type: Improvement
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: 4.2, 5.0
>
>
> Check why we depend on the full core POM anyway; the mini version should be 
> enough for algorithms (we wouldn't need so many exclusions then).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Resolved] (SOLR-4259) Carrot2 dependency should be declared on the mini version, not the core

2013-03-04 Thread Dawid Weiss (JIRA)

 [ 
https://issues.apache.org/jira/browse/SOLR-4259?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dawid Weiss resolved SOLR-4259.
---

Resolution: Fixed

> Carrot2 dependency should be declared on the mini version, not the core
> ---
>
> Key: SOLR-4259
> URL: https://issues.apache.org/jira/browse/SOLR-4259
> Project: Solr
>  Issue Type: Improvement
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: 4.2, 5.0
>
>
> Check why we depend on the full core POM anyway; the mini version should be 
> enough for algorithms (we wouldn't need so many exclusions then).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4259) Carrot2 dependency should be declared on the mini version, not the core

2013-03-04 Thread Commit Tag Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4259?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592083#comment-13592083
 ] 

Commit Tag Bot commented on SOLR-4259:
--

[trunk commit] Dawid Weiss
http://svn.apache.org/viewvc?view=revision&revision=1452220

SOLR-4259: Carrot2 dependency should be declared on the mini version, not the 
core.




> Carrot2 dependency should be declared on the mini version, not the core
> ---
>
> Key: SOLR-4259
> URL: https://issues.apache.org/jira/browse/SOLR-4259
> Project: Solr
>  Issue Type: Improvement
>Reporter: Dawid Weiss
>Assignee: Dawid Weiss
>Priority: Minor
> Fix For: 4.2, 5.0
>
>
> Check why we depend on the full core POM anyway; the mini version should be 
> enough for algorithms (we wouldn't need so many exclusions then).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



[jira] [Commented] (SOLR-4506) [solr4.0.0] many index.{date} dir in replication node

2013-03-04 Thread zhuojunjian (JIRA)

[ 
https://issues.apache.org/jira/browse/SOLR-4506?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592075#comment-13592075
 ] 

zhuojunjian commented on SOLR-4506:
---

hi
do you have any plan for this issue ?
4.1 or other ?

> [solr4.0.0] many index.{date} dir in replication node 
> --
>
> Key: SOLR-4506
> URL: https://issues.apache.org/jira/browse/SOLR-4506
> Project: Solr
>  Issue Type: Bug
>  Components: SolrCloud
>Affects Versions: 4.0
> Environment: the solr4.0 runs on suse11.
> mem:32G
> cpu:16 cores
>Reporter: zhuojunjian
> Fix For: 4.0
>
>   Original Estimate: 12h
>  Remaining Estimate: 12h
>
> in our test,we used solrcloud feature in solr4.0(version detail 
> :4.0.0.2012.10.06.03.04.33).
> the solrcloud configuration is 3 shards and 2 replications each shard.
> we found that there are over than 25 dirs which named index.{date} in one 
> replication node belonging to shard 3. 
> for example:
> index.2013021725864  index.20130218012211880  index.20130218015714713  
> index.20130218023101958  index.20130218030424083  tlog
> index.20130218005648324  index.20130218012751078  index.20130218020141293  
> the issue seems like SOLR-1781. but it is fixed in 4.0-BETA,5.0. 
> so is solr4.0 ? if it is fixed too in solr4.0, why we find the issue again ?
> what can I do?   

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

-
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org



  1   2   >