RE: Retrieving a non-indexed but stored field

2008-11-05 Thread Andrew Nagy
Sorry for the late follow-up.  I am doing this, but get nothing back.

Can anyone replicate this problem?

Andrew

From: Erik Hatcher [EMAIL PROTECTED]
Sent: Tuesday, October 14, 2008 12:36 PM
To: solr-user@lucene.apache.org
Subject: Re: Retrieving a non-indexed but stored field

On Oct 14, 2008, at 12:16 PM, Andrew Nagy wrote:

 Hello - I have the following field:

 field name=url type=string indexed=false stored=true
 multiValued=true/

 However, when I do a search, the url field does not display.  Does
 the field also need to be indexed in order to retrieve the data?

No... but it does need to be requested, either as fl=url, or fl=*

Erik




RE: Retrieving a non-indexed but stored field

2008-11-05 Thread Andrew Nagy
Sorry for the late follow-up.  I am doing this, but get nothing back.

Can anyone replicate this problem?

Andrew

From: Erik Hatcher [EMAIL PROTECTED]
Sent: Tuesday, October 14, 2008 12:36 PM
To: solr-user@lucene.apache.org
Subject: Re: Retrieving a non-indexed but stored field

On Oct 14, 2008, at 12:16 PM, Andrew Nagy wrote:

 Hello - I have the following field:

 field name=url type=string indexed=false stored=true
 multiValued=true/

 However, when I do a search, the url field does not display.  Does
 the field also need to be indexed in order to retrieve the data?

No... but it does need to be requested, either as fl=url, or fl=*

Erik




RE: Retrieving a non-indexed but stored field

2008-11-05 Thread Andrew Nagy
Nope - I made the schema change and then indexed all of my content.

I can confirm that the URL string is included, cause when I change my schema 
back to have both stored and indexed, it shows the URL data in the search 
results.  When I change it to stored and not indexed, no data is returned.

Andrew

From: [EMAIL PROTECTED] [EMAIL PROTECTED] On Behalf Of Yonik Seeley [EMAIL 
PROTECTED]
Sent: Wednesday, November 05, 2008 1:41 PM
To: solr-user@lucene.apache.org
Subject: Re: Retrieving a non-indexed but stored field

On Wed, Nov 5, 2008 at 11:47 AM, Andrew Nagy [EMAIL PROTECTED] wrote:
 Sorry for the late follow-up.  I am doing this, but get nothing back.

Did you change the field to stored in the schema after you added the document?
I've never seen anyone having this problem, so perhaps verify that you
are actually sending the url field when indexing.

-Yonik


odd 500 error

2008-08-27 Thread Andrew Nagy
Hello - I stumbled across an odd error which my intuition is telling me is a 
bug.

Here is my installation:
Solr Specification Version: 1.2.2008.08.13.13.05.16
Lucene Implementation Version: 2.4-dev 685576 - 2008-08-13 10:55:25

I did the following query today:
author:(r*a* AND fisher)

And get the following 500 error:

maxClauseCount is set to 1024

org.apache.lucene.search.BooleanQuery$TooManyClauses: maxClauseCount is set to 
1024
at org.apache.lucene.search.BooleanQuery.add(BooleanQuery.java:165)
at org.apache.lucene.search.BooleanQuery.add(BooleanQuery.java:156)
at 
org.apache.lucene.search.MultiTermQuery.rewrite(MultiTermQuery.java:63)
at org.apache.lucene.search.WildcardQuery.rewrite(WildcardQuery.java:54)
at org.apache.lucene.search.BooleanQuery.rewrite(BooleanQuery.java:385)
at 
org.apache.lucene.search.IndexSearcher.rewrite(IndexSearcher.java:163)
at org.apache.lucene.search.Query.weight(Query.java:94)
at org.apache.lucene.search.Searcher.createWeight(Searcher.java:175)
at org.apache.lucene.search.Searcher.search(Searcher.java:126)
at org.apache.lucene.search.Searcher.search(Searcher.java:105)
at 
org.apache.solr.search.SolrIndexSearcher.getDocListNC(SolrIndexSearcher.java:966)
at 
org.apache.solr.search.SolrIndexSearcher.getDocListC(SolrIndexSearcher.java:838)
at 
org.apache.solr.search.SolrIndexSearcher.search(SolrIndexSearcher.java:269)
at 
org.apache.solr.handler.component.QueryComponent.process(QueryComponent.java:160)
at 
org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:167)
at 
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:135)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:1156)
at 
org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:341)
at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:272)
at 
org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1088)
at 
org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:360)
at 
org.mortbay.jetty.security.SecurityHandler.handle(SecurityHandler.java:216)
at 
org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:181)
at 
org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:729)
at org.mortbay.jetty.webapp.WebAppContext.handle(WebAppContext.java:405)
at 
org.mortbay.jetty.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:206)
at 
org.mortbay.jetty.handler.HandlerCollection.handle(HandlerCollection.java:114)
at 
org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:152)
at org.mortbay.jetty.Server.handle(Server.java:324)
at 
org.mortbay.jetty.HttpConnection.handleRequest(HttpConnection.java:505)
at 
org.mortbay.jetty.HttpConnection$RequestHandler.headerComplete(HttpConnection.java:829)
at org.mortbay.jetty.HttpParser.parseNext(HttpParser.java:513)
at org.mortbay.jetty.HttpParser.parseAvailable(HttpParser.java:211)
at org.mortbay.jetty.HttpConnection.handle(HttpConnection.java:380)
at 
org.mortbay.io.nio.SelectChannelEndPoint.run(SelectChannelEndPoint.java:395)
at 
org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:488)


Thanks
Andrew


RE: NOTICE: multicore.xml changed to solr.xml, format changes as well

2008-08-13 Thread Andrew Nagy
Okay - I found the removal of the default attribute in 
https://svn.apache.org/viewvc/lucene/solr/trunk/src/java/org/apache/solr/core/MultiCore.java?
 r1=606335r2=602003

I will update the documentation on the multicore changes.

Andrew

 -Original Message-
 From: Chris Hostetter [mailto:[EMAIL PROTECTED]
 Sent: Wednesday, August 13, 2008 2:27 AM
 To: solr-user@lucene.apache.org
 Subject: RE: NOTICE: multicore.xml changed to solr.xml, format changes
 as well


 : Chris - thanks for the alert.  Can you please clarify the usage of
 the
 : default attribute that is documented to be used in the core node.
 : Solr-545 has a note about this being removed and it is not shown in
 the
 : new example solr.xml file.

 Any attribute that was in the old example multicore.xml has a
 corrisponding attribute in the example solr.xml ...

 https://svn.apache.org/viewvc/lucene/solr/trunk/example/multicore/solr.
 xml?r1=650331r2=685244

 ...no functionality was changed at all in this commit, it was just
 renamed.

 i don't know anything about a default attribute, other then the fact
 that the previous commit to that file (r650331) had the message
 default
 is no longer a multicore concept

 https://svn.apache.org/viewvc/lucene/solr/trunk/example/multicore/solr.
 xml

 -Hoss



multicore can't find the data directories

2008-08-13 Thread Andrew Nagy
I am trying to setup a multicore system implementation.  I just upgraded to 
today's snapshot and have converted my multicore.xml to solr.war and I also 
changed the xml to match the new schema.  However, now that I have done that, 
Solr is not finding my data directory.  With the use of multicore - does each 
core have its own data directory?

I have the following setup:

solr/core0/conf
solr/core1/conf
solr/solr.xml

When I look at the admin (http://localhost:8080/solr/core0/admin/stats.jsp) it 
shows the following:
readerDir : 
org.apache.lucene.store.FSDirectory@/usr/local/solr/data/indexmailto:org.apache.lucene.store.FSDirectory@/usr/local/solr/data/index

and numDocs is reporting 0.

However I have my index with plenty of records located in solr/core0/data/index

Is Solr looking in the wrong directory, or do I have my index in the wrong 
place?

Thanks
Andrew





RE: multicore can't find the data directories

2008-08-13 Thread Andrew Nagy
Nevermind - sorry.  The data directory in my solrconfig.xml was not changed to 
the correct path.  Now it's alive!

Andrew

 -Original Message-
 From: Andrew Nagy [mailto:[EMAIL PROTECTED]
 Sent: Wednesday, August 13, 2008 12:13 PM
 To: solr-user@lucene.apache.org
 Subject: multicore can't find the data directories

 I am trying to setup a multicore system implementation.  I just
 upgraded to today's snapshot and have converted my multicore.xml to
 solr.war and I also changed the xml to match the new schema.  However,
 now that I have done that, Solr is not finding my data directory.  With
 the use of multicore - does each core have its own data directory?

 I have the following setup:

 solr/core0/conf
 solr/core1/conf
 solr/solr.xml

 When I look at the admin
 (http://localhost:8080/solr/core0/admin/stats.jsp) it shows the
 following:
 readerDir :
 org.apache.lucene.store.FSDirectory@/usr/local/solr/data/indexmailto:o
 rg.apache.lucene.store.FSDirectory@/usr/local/solr/data/index

 and numDocs is reporting 0.

 However I have my index with plenty of records located in
 solr/core0/data/index

 Is Solr looking in the wrong directory, or do I have my index in the
 wrong place?

 Thanks
 Andrew





RE: more multicore fun

2008-08-13 Thread Andrew Nagy
Doug - I had this same problem today.  If you look at my post from earlier 
today you will see the problem.  You will need to adjust the solr.data.dir 
value in the solrconfig.xml.

Maybe this also needs to be changed in the example solrconfig.xml document?

Andrew

 -Original Message-
 From: Doug Steigerwald [mailto:[EMAIL PROTECTED]
 Sent: Wednesday, August 13, 2008 2:53 PM
 To: solr-user@lucene.apache.org
 Subject: more multicore fun

 OK.  Last question for a while (hopefully), but something else with
 multicore seems to be wrong.

 solr persistent=true
cores adminPath=/admin/multicore
  core name=core0 instanceDir=core0/
  core name=core1 instanceDir=core1/
/cores
 /solr

 $ java -jar start.jar
 ...
 INFO: [core0] Opening new SolrCore at solr/core0/, dataDir=./solr/data/
 ...
 INFO: [core1] Opening new SolrCore at solr/core1/, dataDir=./solr/data/
 ...

 The instanceDir seems to be fine, but the dataDir isn't being set
 correctly.  The dataDir is actually example/solr/data instead of
 example/solr/core{0|1}/data.

 http://localhost:8983/solr/admin/multicore shows the exact same path
 to the index for both cores.  Am I missing something that the example
 multicore config doesn't use?

 Thanks.
 Doug


Solr1.3 Freeze

2008-08-12 Thread Andrew Nagy
I read on the Solr 1.3 wiki page that there is a code freeze as of today, is 
this still accurate?  Moreover - does this mean that Solr1.3 will most likely 
ship with Lucene 2.4-dev or is there any plan to wait for lucene 2.4 to be 
released?

I know scheduling questions are annoying, but I am curious as to how to better 
manage a project that uses solr and how releases should be scheduled around 
that.

Thanks!
Andrew


RE: NOTICE: multicore.xml changed to solr.xml, format changes as well

2008-08-12 Thread Andrew Nagy
Chris - thanks for the alert.  Can you please clarify the usage of the default 
attribute that is documented to be used in the core node.  Solr-545 has a 
note about this being removed and it is not shown in the new example solr.xml 
file.

Thanks
Andrew

 -Original Message-
 From: Chris Hostetter [mailto:[EMAIL PROTECTED]
 Sent: Tuesday, August 12, 2008 2:01 PM
 To: solr-user@lucene.apache.org
 Subject: NOTICE: multicore.xml changed to solr.xml, format changes as
 well


 If you've bene using the trunk (and/or nightly builds) and you take
 advantage of the MultiCore features in Solr pelase be aware...

 As of r685244 (committed a few moments ago) Solr no longer looks for a
 multicore.xml file.  It instead looks for a solr.xml file.

 solr.xml supports all of the options that multicore.xml supported,
 however
 they have been tweaked slightly (in some cases renamed, in other
 attributes have been moved from one XML tag to another).

 A detailed example can be seen in example/multicore/solr.xml...

 http://svn.apache.org/viewvc/lucene/solr/trunk/example/multicore/solr.x
 ml?view=markup

 For more information, please see SOLR-689...
 https://issues.apache.org/jira/browse/SOLR-689

 Volunteers to help update the wiki documentation would be appreciated.


 -Hoss



TimeExceededException

2008-07-31 Thread Andrew Nagy
Hello - I am a part of a larger group working on an import tool called 
SolrMarc.  I am running into an error that I'm not sure what is causing it and 
looking for any leads.

I am getting the following exception on the SolrCore constructor:
Exception in thread main java.lang.NoClassDefFoundError: 
org/apache/lucene/search/TimeLimitedCollector$TimeExceededException
at org.apache.solr.core.SolrConfig.init(SolrConfig.java:128)
at org.apache.solr.core.SolrConfig.init(SolrConfig.java:97)
...

Any ideas what might cause this?  I am working from the July 25 nightly 
snapshot.  Could I be missing a jar or something?

Thanks!
Andrew


RE: SpellCheckComponent problems (was: Multiple search components in one handler - ie spellchecker)

2008-07-28 Thread Andrew Nagy
Shalin - yes the allfields field exists in my schema.xml file.  It is a field 
that has all of the text from all of the fields concatenated together into one 
field.

My spellCheckIndexDir is created and has 2 segment files, but I think the index 
is empty.  When I initiate the 1st spellcheck.build=true ... the results load 
immediately ... I would imagine some time delay as it builds the index.

Any other ideas?

Andrew

 -Original Message-
 From: Shalin Shekhar Mangar [mailto:[EMAIL PROTECTED]
 Sent: Friday, July 25, 2008 3:35 PM
 To: solr-user@lucene.apache.org
 Subject: Re: Multiple search components in one handler - ie
 spellchecker

 On Sat, Jul 26, 2008 at 12:37 AM, Andrew Nagy
 [EMAIL PROTECTED]
 wrote:

  Exactly - however the spellcheck component is not working for my
 setup.
   The spelling suggestions never show in the response.  I think I have
 the
  solrconfig setup incorrectly.  Also my solr/data/spell index that is
 created
  is empty.  Something is not configured correctly, any ideas?
 
  Andrew
  
  From: Geoffrey Young [EMAIL PROTECTED]
  Sent: Friday, July 25, 2008 3:04 PM
  To: solr-user@lucene.apache.org
  Subject: Re: Multiple search components in one handler - ie
 spellchecker
 
  Andrew Nagy wrote:
   Thanks for getting back to me Geoff.  Although, that is pretty much
   what I have.  Maybe if I show my solrconfig someone might be able
 to
   point out what I have incorrect?  The problem is that nothing
 related
   to the spelling options are show in the results, just the normal
   expected search results.
 
  right.  the spellcheck component does not issue a separate query
 *after*
  running the spellcheck, it merely offers suggestions in parallel with
  your existing query.
 
  the results are more like
 
below are the results for $query.  did you mean $suggestions?
 
  HTH
 
  --Geoff
 
 
 
 Is the allfields in your spell checker configuration in your
 schema.xml? Can
 you see the spellcheckIndexDir created inside the Solr's data
 directory?

 --
 Regards,
 Shalin Shekhar Mangar.


RE: SpellCheckComponent problems (was: Multiple search components in one handler - ie spellchecker)

2008-07-28 Thread Andrew Nagy
 -Original Message-
 From: Shalin Shekhar Mangar [mailto:[EMAIL PROTECTED]
 Sent: Monday, July 28, 2008 10:09 AM
 To: solr-user@lucene.apache.org
 Subject: Re: SpellCheckComponent problems (was: Multiple search
 components in one handler - ie spellchecker)

 Can you show us the query you are issuing? Make sure you add
 spellcheck=true
 to the query as a parameter to turn on spell checking.

http://localhost:8080/solr/select?q=*:*spellcheck=truespellcheck.q=scandanavaspellcheck.build=true

Shows this:
response
lst name=responseHeader
int name=status0/int
int name=QTime73/int
/lst
result name=response numFound=7650 start=0
...
/result
/response

Andrew


RE: SpellCheckComponent problems (was: Multiple search components in one handler - ie spellchecker)

2008-07-28 Thread Andrew Nagy
I was just reviewing the solr logs and I noticed the following:

Jul 28, 2008 11:52:01 AM org.apache.solr.common.SolrException log
SEVERE: org.apache.solr.common.SolrException: Error loading class 
'org.apache.solr.handler.component.SpellCheckComponent'

It looks like the SpellCheckComponent is not getting loaded.  What could cause 
this?  Im running the july25 nightly build.

Here is a list of the libs from my /tmp/jetty/webapp/WEB-INF/lib dir:
-rw-r--r--  1 root root  84199 Jul 25 08:14 apache-solr-common-nightly.jar
-rw-r--r--  1 root root 889903 Jul 25 08:14 apache-solr-nightly.jar
-rw-r--r--  1 root root  46725 May 10  2007 commons-codec-1.3.jar
-rw-r--r--  1 root root  22017 Jan  6  2008 commons-csv-1.0-SNAPSHOT-r609327.jar
-rw-r--r--  1 root root  53082 Mar  1  2007 commons-fileupload-1.2.jar
-rw-r--r--  1 root root 305001 Sep 11  2007 commons-httpclient-3.1.jar
-rw-r--r--  1 root root  83613 Jun 15  2007 commons-io-1.3.1.jar
-rw-r--r--  1 root root  38015 Jun 14  2007 commons-logging-1.0.4.jar
-rw-r--r--  1 root root 249154 Sep 21  2007 junit-4.3.jar
-rw-r--r--  1 root root 115101 Jun 19 13:46 lucene-analyzers-2.4-dev.jar
-rw-r--r--  1 root root 730352 Jun 19 13:46 lucene-core-2.4-dev.jar
-rw-r--r--  1 root root  87390 Jun 19 13:46 lucene-highlighter-2.4-dev.jar
-rw-r--r--  1 root root  32693 Jun 19 13:46 lucene-queries-2.4-dev.jar
-rw-r--r--  1 root root  91029 Jun 19 13:46 lucene-snowball-2.4-dev.jar
-rw-r--r--  1 root root  18422 Jun 19 13:46 lucene-spellchecker-2.4-dev.jar
-rw-r--r--  1 root root 179348 Jun 14  2007 stax-1.2.0-dev.jar
-rw-r--r--  1 root root  25863 Jun 14  2007 stax-api-1.0.jar
-rw-r--r--  1 root root 128475 Jun 14  2007 stax-utils.jar

could I be missing a jar?

Thanks
Andrew

 -Original Message-
 From: Shalin Shekhar Mangar [mailto:[EMAIL PROTECTED]
 Sent: Monday, July 28, 2008 11:24 AM
 To: solr-user@lucene.apache.org
 Subject: Re: SpellCheckComponent problems (was: Multiple search
 components in one handler - ie spellchecker)

 Hi Andrew,

 Your configuration which you specified in the earlier thread looks
 fine.
 Your query is also ok. The complete lack of spell check results in the
 response you pasted suggests that the SpellCheckComponent is not added
 to
 the SearchHandler's list of components.

 Can you check your solrconfig.xml again? I'm sorry but it doesn't seem
 like
 a problem with the spell checker itself. Also check if there are any
 exceptions in the Solr log/console.

 On Mon, Jul 28, 2008 at 8:32 PM, Andrew Nagy
 [EMAIL PROTECTED]wrote:

   -Original Message-
   From: Shalin Shekhar Mangar [mailto:[EMAIL PROTECTED]
   Sent: Monday, July 28, 2008 10:09 AM
   To: solr-user@lucene.apache.org
   Subject: Re: SpellCheckComponent problems (was: Multiple search
   components in one handler - ie spellchecker)
  
   Can you show us the query you are issuing? Make sure you add
   spellcheck=true
   to the query as a parameter to turn on spell checking.
 
 
 
 http://localhost:8080/solr/select?q=*:*spellcheck=truespellcheck.q=sc
 andanavaspellcheck.build=true
 
  Shows this:
  response
  lst name=responseHeader
  int name=status0/int
  int name=QTime73/int
  /lst
  result name=response numFound=7650 start=0
  ...
  /result
  /response
 
  Andrew
 



 --
 Regards,
 Shalin Shekhar Mangar.


RE: SpellCheckComponent problems (was: Multiple search components in one handler - ie spellchecker)

2008-07-28 Thread Andrew Nagy
Hmm ... sorry, that was the output of a java program that uses solr that I ran 
and noticed the error.  That error doesn't happen when I start solr.  Sorry for 
the confusion.

I just changed my schema to have a dedicated field for spelling called 
spelling and I created a new field type for the spellcheck component called 
textSpell.
Here is the segment of my solrconfig.xml:

searchComponent name=spellcheck 
class=org.apache.solr.handler.component.SpellCheckComponent
  lst name=spellchecker
str name=fieldspelling/str
str name=accuracy0.7/str !-- 0.5 is the default setting --
str name=spellcheckIndexDir./spellchecker/str
  /lst
  str name=queryAnalyzerFieldTypetextSpell/str
/searchComponent
queryConverter name=queryConverter 
class=org.apache.solr.spelling.SpellingQueryConverter/
requestHandler name=/search 
class=org.apache.solr.handler.component.SearchHandler
  lst name=defaults
str name=echoParamsexplicit/str
  /lst
  arr name=last-components
strspellcheck/str
!--strelevator/str--
  /arr
/requestHandler

I will need to reindex my documents again - I will check to see if that has any 
effect on my problem.

Andrew


 -Original Message-
 From: Shalin Shekhar Mangar [mailto:[EMAIL PROTECTED]
 Sent: Monday, July 28, 2008 12:07 PM
 To: solr-user@lucene.apache.org
 Subject: Re: SpellCheckComponent problems (was: Multiple search
 components in one handler - ie spellchecker)

 No, SpellCheckComponent was in the nightly long before July 25. There
 must
 be a stack trace after that error message. Can you post that?

 On Mon, Jul 28, 2008 at 9:26 PM, Andrew Nagy
 [EMAIL PROTECTED]wrote:

  I was just reviewing the solr logs and I noticed the following:
 
  Jul 28, 2008 11:52:01 AM org.apache.solr.common.SolrException log
  SEVERE: org.apache.solr.common.SolrException: Error loading class
  'org.apache.solr.handler.component.SpellCheckComponent'
 
  It looks like the SpellCheckComponent is not getting loaded.  What
 could
  cause this?  Im running the july25 nightly build.
 
  Here is a list of the libs from my /tmp/jetty/webapp/WEB-INF/lib dir:
  -rw-r--r--  1 root root  84199 Jul 25 08:14 apache-solr-common-
 nightly.jar
  -rw-r--r--  1 root root 889903 Jul 25 08:14 apache-solr-nightly.jar
  -rw-r--r--  1 root root  46725 May 10  2007 commons-codec-1.3.jar
  -rw-r--r--  1 root root  22017 Jan  6  2008
  commons-csv-1.0-SNAPSHOT-r609327.jar
  -rw-r--r--  1 root root  53082 Mar  1  2007 commons-fileupload-
 1.2.jar
  -rw-r--r--  1 root root 305001 Sep 11  2007 commons-httpclient-
 3.1.jar
  -rw-r--r--  1 root root  83613 Jun 15  2007 commons-io-1.3.1.jar
  -rw-r--r--  1 root root  38015 Jun 14  2007 commons-logging-1.0.4.jar
  -rw-r--r--  1 root root 249154 Sep 21  2007 junit-4.3.jar
  -rw-r--r--  1 root root 115101 Jun 19 13:46 lucene-analyzers-2.4-
 dev.jar
  -rw-r--r--  1 root root 730352 Jun 19 13:46 lucene-core-2.4-dev.jar
  -rw-r--r--  1 root root  87390 Jun 19 13:46 lucene-highlighter-2.4-
 dev.jar
  -rw-r--r--  1 root root  32693 Jun 19 13:46 lucene-queries-2.4-
 dev.jar
  -rw-r--r--  1 root root  91029 Jun 19 13:46 lucene-snowball-2.4-
 dev.jar
  -rw-r--r--  1 root root  18422 Jun 19 13:46 lucene-spellchecker-2.4-
 dev.jar
  -rw-r--r--  1 root root 179348 Jun 14  2007 stax-1.2.0-dev.jar
  -rw-r--r--  1 root root  25863 Jun 14  2007 stax-api-1.0.jar
  -rw-r--r--  1 root root 128475 Jun 14  2007 stax-utils.jar
 
  could I be missing a jar?
 
  Thanks
  Andrew
 
   -Original Message-
   From: Shalin Shekhar Mangar [mailto:[EMAIL PROTECTED]
   Sent: Monday, July 28, 2008 11:24 AM
   To: solr-user@lucene.apache.org
   Subject: Re: SpellCheckComponent problems (was: Multiple search
   components in one handler - ie spellchecker)
  
   Hi Andrew,
  
   Your configuration which you specified in the earlier thread looks
   fine.
   Your query is also ok. The complete lack of spell check results in
 the
   response you pasted suggests that the SpellCheckComponent is not
 added
   to
   the SearchHandler's list of components.
  
   Can you check your solrconfig.xml again? I'm sorry but it doesn't
 seem
   like
   a problem with the spell checker itself. Also check if there are
 any
   exceptions in the Solr log/console.
  
   On Mon, Jul 28, 2008 at 8:32 PM, Andrew Nagy
   [EMAIL PROTECTED]wrote:
  
 -Original Message-
 From: Shalin Shekhar Mangar [mailto:[EMAIL PROTECTED]
 Sent: Monday, July 28, 2008 10:09 AM
 To: solr-user@lucene.apache.org
 Subject: Re: SpellCheckComponent problems (was: Multiple search
 components in one handler - ie spellchecker)

 Can you show us the query you are issuing? Make sure you add
 spellcheck=true
 to the query as a parameter to turn on spell checking.
   
   
   
  
 http://localhost:8080/solr/select?q=*:*spellcheck=truespellcheck.q=sc
   andanavaspellcheck.build=true
   
Shows this:
response
lst name=responseHeader
int name=status0/int
int name=QTime73/int
/lst
result

RE: SpellCheckComponent problems (was: Multiple search components in one handler - ie spellchecker)

2008-07-28 Thread Andrew Nagy
Well I will include the stack trace for the aforementioned error:

Jul 28, 2008 12:20:17 PM org.apache.solr.common.SolrException log
SEVERE: org.apache.solr.common.SolrException: Error loading class 
'org.apache.solr.handler.component.SpellCheckComponent'
at 
org.apache.solr.core.SolrResourceLoader.findClass(SolrResourceLoader.java:227)
at 
org.apache.solr.core.SolrResourceLoader.newInstance(SolrResourceLoader.java:232)
at 
org.apache.solr.util.plugin.AbstractPluginLoader.create(AbstractPluginLoader.java:83)
at 
org.apache.solr.util.plugin.AbstractPluginLoader.load(AbstractPluginLoader.java:140)
at org.apache.solr.core.SolrCore.loadSearchComponents(SolrCore.java:565)
at org.apache.solr.core.SolrCore.init(SolrCore.java:371)
at org.solrmarc.marc.MarcImporter.init(MarcImporter.java:95)
at org.solrmarc.marc.MarcImporter.main(MarcImporter.java:559)
Caused by: java.lang.ClassNotFoundException: 
org.apache.solr.handler.component.SpellCheckComponent
at java.net.URLClassLoader$1.run(URLClassLoader.java:200)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:188)
at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
at java.net.FactoryURLClassLoader.loadClass(URLClassLoader.java:580)
at java.lang.ClassLoader.loadClass(ClassLoader.java:251)
at java.lang.ClassLoader.loadClassInternal(ClassLoader.java:319)
at java.lang.Class.forName0(Native Method)
at java.lang.Class.forName(Class.java:242)
at 
org.apache.solr.core.SolrResourceLoader.findClass(SolrResourceLoader.java:211)
... 7 more

The line 95 of MarcImporter.java (the solr import program I am using) is the 
instantiation of SolrCore.  So maybe somehow the spellCheckComponent is not 
getting loaded?

This is the error output I get thrown by instantiating SolrCore:
org.apache.solr.common.SolrException: Unknown Search Component: spellcheck
at org.apache.solr.core.SolrCore.getSearchComponent(SolrCore.java:597)
at 
org.apache.solr.handler.component.SearchHandler.inform(SearchHandler.java:107)
at 
org.apache.solr.core.SolrResourceLoader.inform(SolrResourceLoader.java:264)
at org.apache.solr.core.SolrCore.init(SolrCore.java:398)
at org.solrmarc.marc.MarcImporter.init(MarcImporter.java:95)
at org.solrmarc.marc.MarcImporter.main(MarcImporter.java:559)

Andrew

 -Original Message-
 From: Shalin Shekhar Mangar [mailto:[EMAIL PROTECTED]
 Sent: Monday, July 28, 2008 12:07 PM
 To: solr-user@lucene.apache.org
 Subject: Re: SpellCheckComponent problems (was: Multiple search
 components in one handler - ie spellchecker)

 No, SpellCheckComponent was in the nightly long before July 25. There
 must
 be a stack trace after that error message. Can you post that?

 On Mon, Jul 28, 2008 at 9:26 PM, Andrew Nagy
 [EMAIL PROTECTED]wrote:

  I was just reviewing the solr logs and I noticed the following:
 
  Jul 28, 2008 11:52:01 AM org.apache.solr.common.SolrException log
  SEVERE: org.apache.solr.common.SolrException: Error loading class
  'org.apache.solr.handler.component.SpellCheckComponent'
 
  It looks like the SpellCheckComponent is not getting loaded.  What
 could
  cause this?  Im running the july25 nightly build.
 
  Here is a list of the libs from my /tmp/jetty/webapp/WEB-INF/lib dir:
  -rw-r--r--  1 root root  84199 Jul 25 08:14 apache-solr-common-
 nightly.jar
  -rw-r--r--  1 root root 889903 Jul 25 08:14 apache-solr-nightly.jar
  -rw-r--r--  1 root root  46725 May 10  2007 commons-codec-1.3.jar
  -rw-r--r--  1 root root  22017 Jan  6  2008
  commons-csv-1.0-SNAPSHOT-r609327.jar
  -rw-r--r--  1 root root  53082 Mar  1  2007 commons-fileupload-
 1.2.jar
  -rw-r--r--  1 root root 305001 Sep 11  2007 commons-httpclient-
 3.1.jar
  -rw-r--r--  1 root root  83613 Jun 15  2007 commons-io-1.3.1.jar
  -rw-r--r--  1 root root  38015 Jun 14  2007 commons-logging-1.0.4.jar
  -rw-r--r--  1 root root 249154 Sep 21  2007 junit-4.3.jar
  -rw-r--r--  1 root root 115101 Jun 19 13:46 lucene-analyzers-2.4-
 dev.jar
  -rw-r--r--  1 root root 730352 Jun 19 13:46 lucene-core-2.4-dev.jar
  -rw-r--r--  1 root root  87390 Jun 19 13:46 lucene-highlighter-2.4-
 dev.jar
  -rw-r--r--  1 root root  32693 Jun 19 13:46 lucene-queries-2.4-
 dev.jar
  -rw-r--r--  1 root root  91029 Jun 19 13:46 lucene-snowball-2.4-
 dev.jar
  -rw-r--r--  1 root root  18422 Jun 19 13:46 lucene-spellchecker-2.4-
 dev.jar
  -rw-r--r--  1 root root 179348 Jun 14  2007 stax-1.2.0-dev.jar
  -rw-r--r--  1 root root  25863 Jun 14  2007 stax-api-1.0.jar
  -rw-r--r--  1 root root 128475 Jun 14  2007 stax-utils.jar
 
  could I be missing a jar?
 
  Thanks
  Andrew
 
   -Original Message-
   From: Shalin Shekhar Mangar [mailto:[EMAIL PROTECTED]
   Sent: Monday, July 28, 2008 11:24 AM
   To: solr-user@lucene.apache.org
   Subject: Re: SpellCheckComponent

RE: SpellCheckComponent problems (was: Multiple search components in one handler - ie spellchecker)

2008-07-28 Thread Andrew Nagy
 -Original Message-
 From: Shalin Shekhar Mangar [mailto:[EMAIL PROTECTED]
 Sent: Monday, July 28, 2008 12:38 PM
 To: solr-user@lucene.apache.org
 Subject: Re: SpellCheckComponent problems (was: Multiple search
 components in one handler - ie spellchecker)

 Well that means the nightly solr jar you are using is older than you
 think
 it is. Try running solr normally without the program and see if you can
 get
 it working.

Well my import program has an older copy of the solr libs ...  so we can ignore 
that problem.

However my problem still stands when I run solr normally from my July25 
snapshot.  There are no errors - and no output to the solr logs when I post a 
query.

Have you or anyone been able to successfully add the spellcheckcomponent to the 
default select searchhandler?


Thanks
Andrew


Multiple search components in one handler - ie spellchecker

2008-07-25 Thread Andrew Nagy
Hello - I am attempting to add the spellCheck component in my search 
requesthandler so when a users does a search, they get the results and spelling 
corrections all in one query just like the way the facets work.

I am having some trouble accomplishing this - can anyone point me to 
documentation (other than http://wiki.apache.org/solr/SpellCheckComponent) on 
how to do this or an example solrconfig that would do this correctly?

My guess is there is something preventing this from happening since the 
requesthandler tag is loading the search component and not the spellchecker 
component?

Thanks and sorry if my message is a bit vague

Andrew



RE: Multiple search components in one handler - ie spellchecker

2008-07-25 Thread Andrew Nagy
Thanks for getting back to me Geoff.  Although, that is pretty much what I 
have.  Maybe if I show my solrconfig someone might be able to point out what I 
have incorrect?  The problem is that nothing related to the spelling options 
are show in the results, just the normal expected search results.  Here is my 
query:
http://localhost:8080/solr/select?q=*:*rows=1spellcheck=truespellcheck.q=scandanavaspellcheck.build=true

searchComponent name=spellcheck 
class=org.apache.solr.handler.component.SpellCheckComponent
  lst name=spellchecker
str name=fieldallfields/str
str name=accuracy0.7/str !-- 0.5 is the default setting --
str name=spellcheckIndexDir./spellchecker/str
  /lst
  str name=queryAnalyzerFieldTypetext/str
/searchComponent
queryConverter name=queryConverter 
class=org.apache.solr.spelling.SpellingQueryConverter/
requestHandler name=/search 
class=org.apache.solr.handler.component.SearchHandler
  lst name=defaults
str name=echoParamsexplicit/str
  /lst
  arr name=last-components
strspellcheck/str
!--strelevator/str--
  /arr
/requestHandler


From: Geoffrey Young [EMAIL PROTECTED]
Sent: Friday, July 25, 2008 2:13 PM
To: solr-user@lucene.apache.org
Subject: Re: Multiple search components in one handler - ie spellchecker

Andrew Nagy wrote:
 Hello - I am attempting to add the spellCheck component in my
 search requesthandler so when a users does a search, they get the
 results and spelling corrections all in one query just like the way
 the facets work.

 I am having some trouble accomplishing this - can anyone point me to
 documentation (other than
 http://wiki.apache.org/solr/SpellCheckComponent) on how to do this or
 an example solrconfig that would do this correctly?

http://mail-archives.apache.org/mod_mbox/lucene-solr-user/200806.mbox/[EMAIL 
PROTECTED]

in general, just add the

   arr name=last-components
 strspellcheck/str
   /arr

bit to your existing handler after following setup in the twiki docs.

you can ignore the part about the exceptions, as that has been fixed in
trunk.

HTH

--Geoff


RE: Multiple search components in one handler - ie spellchecker

2008-07-25 Thread Andrew Nagy
Exactly - however the spellcheck component is not working for my setup.  The 
spelling suggestions never show in the response.  I think I have the solrconfig 
setup incorrectly.  Also my solr/data/spell index that is created is empty.  
Something is not configured correctly, any ideas?

Andrew

From: Geoffrey Young [EMAIL PROTECTED]
Sent: Friday, July 25, 2008 3:04 PM
To: solr-user@lucene.apache.org
Subject: Re: Multiple search components in one handler - ie spellchecker

Andrew Nagy wrote:
 Thanks for getting back to me Geoff.  Although, that is pretty much
 what I have.  Maybe if I show my solrconfig someone might be able to
 point out what I have incorrect?  The problem is that nothing related
 to the spelling options are show in the results, just the normal
 expected search results.

right.  the spellcheck component does not issue a separate query *after*
running the spellcheck, it merely offers suggestions in parallel with
your existing query.

the results are more like

   below are the results for $query.  did you mean $suggestions?

HTH

--Geoff




RE: POSTing repeated fields to Solr

2008-06-05 Thread Andrew Nagy
 On 4-Jun-08, at 2:22 PM, Andrew Nagy wrote:

  Hello - I was wondering if there is a work around with POSTing
  repeated fields to Solr.  I am using Jetty as my container with Solr
  1.2.
 
  I tried something like:
 
 http://localhost:8080/solr/select/?q=author:(smith)rows=0start=0face
 t=truefacet.mincount=1facet.limit=10facet.field=authorlastfacet.fie
 ld=authorfirst
 
  I am only getting back facets from the last facet.field.  With a GET
  request I get back all of the facet fields.  I am assuming this is a
  limitation with Jetty?  How are others doing this?

 It could be, but I would be surprised.  How are you constructing the
 POST request  (you listed a url, which only makes sense as a GET)?

I was just listing the GET representation of the URL.  I am using the 
PEAR::HTTP_Request library for PHP.  If you don't think this is a Jetty 
limitation, then it is probably something with the library that I am using.  I 
will investigate there 1st.

Thanks!
Andrew


POSTing repeated fields to Solr

2008-06-04 Thread Andrew Nagy
Hello - I was wondering if there is a work around with POSTing repeated fields 
to Solr.  I am using Jetty as my container with Solr 1.2.

I tried something like:
http://localhost:8080/solr/select/?q=author:(smith)rows=0start=0facet=truefacet.mincount=1facet.limit=10facet.field=authorlastfacet.field=authorfirst

I am only getting back facets from the last facet.field.  With a GET request I 
get back all of the facet fields.  I am assuming this is a limitation with 
Jetty?  How are others doing this?

Thanks
Andrew




Limit on # of fields to query on

2008-03-13 Thread Andrew Nagy
I am testing around with a new feature in my system that uses Solr and I am 
testing a query that has a search on the same field OR'd together over 150 
times.  I know this sounds pretty ridiculous and as I said I am just playing 
around.  However Solr just returns a blank page and doesn't process the query.  
Is there a limit to the number of fields that you can have in your query?

My example is to do some inventory checking with something like:
id:1 or id:5 or id:7 or id:8 ... and so on for roughly 176 times (which I think 
would be the worst case scenario in my catalog)

Thanks!
Andrew


RE: Re:

2007-12-02 Thread Andrew Nagy
Sorry for not explaining my self clearly: I have header=true as you can see 
from the curl command and there is a header line in the csv file.

And sorry for the missing subject line.

Andrew

From: Ryan McKinley [EMAIL PROTECTED]
Sent: Sunday, December 02, 2007 5:15 PM
To: solr-user@lucene.apache.org
Subject: Re:

Andrew Nagy wrote:
 Hello - I am trying out the CSV importer and am curious with an error that I 
 am consistently running into.  What am I doing incorrectly here?  I am 
 importing a pipe delimited CSV file with quotes encapsulation.

 Thanks
 Andrew

 curl 
 http://localhost:8080/solr/update/csv?header=true%26seperator=%7C%26encapsulator=%22%26commit=true%26stream.file=import/homes.csv

 html
 head
 meta http-equiv=Content-Type content=text/html; charset=ISO-8859-1/
 titleError 400 /title
 /head
 bodyh2HTTP ERROR: 400/h2preCSVLoader: must specify 
 fieldnames=lt;fieldsgt;* or header=true/pre
 pRequestURI=/solr/update/csv/ppismalla 
 href=http://jetty.mortbay.org/;Powered by Jetty:///a/small/i/pbr/


take a look at the error:

CSVLoader: must specify fieldnames=lt;fieldsgt;* or header=true

either list the fieldnames or use the 1st line as the field names
(header=true)

ryan






[no subject]

2007-12-02 Thread Andrew Nagy
Hello - I am trying out the CSV importer and am curious with an error that I am 
consistently running into.  What am I doing incorrectly here?  I am importing a 
pipe delimited CSV file with quotes encapsulation.

Thanks
Andrew

curl 
http://localhost:8080/solr/update/csv?header=true%26seperator=%7C%26encapsulator=%22%26commit=true%26stream.file=import/homes.csv

html
head
meta http-equiv=Content-Type content=text/html; charset=ISO-8859-1/
titleError 400 /title
/head
bodyh2HTTP ERROR: 400/h2preCSVLoader: must specify 
fieldnames=lt;fieldsgt;* or header=true/pre
pRequestURI=/solr/update/csv/ppismalla 
href=http://jetty.mortbay.org/;Powered by Jetty:///a/small/i/pbr/


RE: Re:

2007-12-02 Thread Andrew Nagy
 On Dec 2, 2007, at 5:43 PM, Ryan McKinley wrote:


 try \ rather then %26


 or just put quotes around the whole url. I think curl does the right thing 
 here.

I tried all the methods: converting  to %26, converting  to \ and 
encapsulating the url with quotes.  All give the same error.

curl 
http://localhost:8080/solr/update/csv?header=true\seperator=%7C\encapsulator=%22\commit=true\stream.file=import/homes.csv
html
head
meta http-equiv=Content-Type content=text/html; charset=ISO-8859-1/
titleError 400 /title
/head
bodyh2HTTP ERROR: 400/h2preCSVLoader: must specify 
fieldnames=lt;fieldsgt;* or header=true/pre
pRequestURI=/solr/update/csv/ppismalla 
href=http://jetty.mortbay.org/;Powered by Jetty:///a/small/i/pbr/

Any other ideas?  How do other do this?

Thanks
Andrew


RE: RE: Re:

2007-12-02 Thread Andrew Nagy
Ugh ... I shouldn't be coding on a sunday night - especially after the eagles 
lost again!

I spelled separator correctly this time :) - But still no luck.

curl 
'http://localhost:8080/solr/update/csv?header=trueseparator=%7Cencapsulator=%22commit=truestream.file=import/homes.csv'
 -H 'Content-type:text/plain; charset=utf-8'
html
head
meta http-equiv=Content-Type content=text/html; charset=ISO-8859-1/
titleError 400 /title
/head
bodyh2HTTP ERROR: 400/h2preCSVLoader: must specify 
fieldnames=lt;fieldsgt;* or header=true/pre
pRequestURI=/solr/update/csv/ppismalla 
href=http://jetty.mortbay.org/;Powered by Jetty:///a/small/i/pbr/


RE: Re:

2007-12-02 Thread Andrew Nagy
Ryan, i didn't know there was a debugger - this could come in handy for other 
things.  Thanks!

I tried it out and it looks like everything is being parsed correctly when 
passing the url in quotes:

curl 
http://localhost:8080/solr/debug/dump?header=trueseparator=%7Cencapsulator=%22commit=true;
 --data-binary @homes.csv -H 'Content-type:text/plain; charset=utf-8'
?xml version=1.0 encoding=UTF-8?
response
lst name=responseHeaderint name=status0/intint 
name=QTime0/intstr 
name=handlerorg.apache.solr.handler.DumpRequestHandler/strlst 
name=paramsstr name=headertrue/strstr name=separator|/strstr 
name=committrue/strstr name=encapsulator/str/lst/lstlst 
name=paramsstr name=echoHandlertrue/strstr 
name=echoParamsexplicit/strstr name=headertrue/strstr 
name=separator|/strstr name=committrue/strstr 
name=encapsulator/str/lstlst name=contextstr 
name=path/debug/dump/str/lst
/response

I tried again with update/csv and same error occurs:
curl 
http://localhost:8080/solr/update/csv?header=trueseparator=%7Cencapsulator=%22commit=true;
 --data-binary @homes.csv -H 'Content-type:text/plain; charset=utf-8'
html
head
meta http-equiv=Content-Type content=text/html; charset=ISO-8859-1/
titleError 400 /title
/head
bodyh2HTTP ERROR: 400/h2preCSVLoader: must specify 
fieldnames=lt;fieldsgt;* or header=true/pre
pRequestURI=/solr/update/csv/ppismalla 
href=http://jetty.mortbay.org/;Powered by Jetty:///a/small/i/pbr/


RE: Re:

2007-12-02 Thread Andrew Nagy
Perfect!  That did it.

Thanks for debugging this with me :)

Andrew

From: Ryan McKinley [EMAIL PROTECTED]
Sent: Sunday, December 02, 2007 7:10 PM
To: solr-user@lucene.apache.org
Subject: Re:

hymmm - give it a try without specifying header=true

Looks like if you don't specify header=true, it defaults to true - but
if you do, it throws an error.

I think there may be a bug... Yonik, should line 243 be:

   } else if (!hasHeader) {
  ^!!!
ryan


Andrew Nagy wrote:
 Ryan, i didn't know there was a debugger - this could come in handy for other 
 things.  Thanks!

 I tried it out and it looks like everything is being parsed correctly when 
 passing the url in quotes:

 curl 
 http://localhost:8080/solr/debug/dump?header=trueseparator=%7Cencapsulator=%22commit=true;
  --data-binary @homes.csv -H 'Content-type:text/plain; charset=utf-8'
 ?xml version=1.0 encoding=UTF-8?
 response
 lst name=responseHeaderint name=status0/intint 
 name=QTime0/intstr 
 name=handlerorg.apache.solr.handler.DumpRequestHandler/strlst 
 name=paramsstr name=headertrue/strstr name=separator|/strstr 
 name=committrue/strstr name=encapsulator/str/lst/lstlst 
 name=paramsstr name=echoHandlertrue/strstr 
 name=echoParamsexplicit/strstr name=headertrue/strstr 
 name=separator|/strstr name=committrue/strstr 
 name=encapsulator/str/lstlst name=contextstr 
 name=path/debug/dump/str/lst
 /response

 I tried again with update/csv and same error occurs:
 curl 
 http://localhost:8080/solr/update/csv?header=trueseparator=%7Cencapsulator=%22commit=true;
  --data-binary @homes.csv -H 'Content-type:text/plain; charset=utf-8'
 html
 head
 meta http-equiv=Content-Type content=text/html; charset=ISO-8859-1/
 titleError 400 /title
 /head
 bodyh2HTTP ERROR: 400/h2preCSVLoader: must specify 
 fieldnames=lt;fieldsgt;* or header=true/pre
 pRequestURI=/solr/update/csv/ppismalla 
 href=http://jetty.mortbay.org/;Powered by Jetty:///a/small/i/pbr/




performance questions

2007-08-30 Thread Andrew Nagy
Here are a few SOLR performance questions:

1. I have noticed with 500,000+ records that my facets run quite fast regarding 
my dataset when there is a large number of matches, but on a small result set 
(say 10 - 50) the facet queries become very slow.  Any suggestions as to how to 
improve this?

2. Someone asked me if SOLR utilizes anything like a stored procedure to make 
queries faster.  Does SOLR support anything such as this?

Thanks
Andrew


Apache web server logs in solr

2007-08-22 Thread Andrew Nagy
Hello, I was thinking that solr - with its built in faceting - would make for a 
great apache log file storage system.  I was wondering if anyone knows of any 
module or library for apache to write log files directly to solr or to a lucene 
index?

Thanks
Andrew


Facet Field Limits

2007-07-12 Thread Andrew Nagy
Hello, I would like to generate a list of facets, let's say on 5 fields.  I 
have the facet limit set to 5 so that for each of the 5 fields there will only 
by up to 5 values.

My question is: Is there a way to change the limit per field?  Let's say on 
facet 2 I would like to display 10 values instead of 5 like the other facets.

Thanks!
Andrew


RE: compile error with SOLR 69 MoreLikeThis patch

2007-05-24 Thread Andrew Nagy
That did the trick.  However, now I am trying to apply the patch to a fresh 
copy of solr on another server.  Whenever I do so, I get the following error 
from the patch command:

patching file src/java/org/apache/solr/request/StandardRequestHandler.java
Hunk #1 succeeded at 23 with fuzz 2 (offset -1 lines).
Hunk #2 FAILED at 31.
Hunk #3 succeeded at 146 with fuzz 2 (offset 13 lines).
1 out of 3 hunks FAILED -- saving rejects to file 
src/java/org/apache/solr/request/StandardRequestHandler.java.rej
(Stripping trailing CRs from patch.)
patching file src/java/org/apache/solr/handler/MoreLikeThisHandler.java
(Stripping trailing CRs from patch.)
patching file src/java/org/apache/solr/util/MoreLikeThisParams.java
(Stripping trailing CRs from patch.)
patching file example/solr/conf/solrconfig.xml
Hunk #1 FAILED at 385.
1 out of 1 hunk FAILED -- saving rejects to file 
example/solr/conf/solrconfig.xml.rej


Any ideas on how to get passed this?

Andrew

From: Brian Whitman [EMAIL PROTECTED]
Sent: Wednesday, May 16, 2007 11:14 AM
To: solr-user
Subject: Re: compile error with SOLR 69 MoreLikeThis patch

Change it to DEFALT or change the spelling error in the Lucene version.

On May 16, 2007, at 12:13 PM, Andrew Nagy wrote:

 I downloaded and patched my solr source with the latest solr69
 patch and whenever I run ant I get an error:

 [javac] /office/src/apache-solr-nightly/src/java/org/apache/solr/
 handler/MoreLikeThisHandler.java:145: cannot find symbol
 [javac] symbol  : variable DEFAULT_MIN_DOC_FREQ
 [javac] location: class org.apache.lucene.search.similar.MoreLikeThis
 [javac]   mlt.setMinDocFreq(params.getInt
 (MoreLikeThisParams.MIN_DOC_FREQ,
 MoreLikeThis.DEFAULT_MIN_DOC_FREQ));
 [javac]
   ^


 Did I miss something, or maybe I have a non-compatible version of
 SOLR?

 Thanks
 Andrew

--
http://variogr.am/
[EMAIL PROTECTED]






RE: compile error with SOLR 69 MoreLikeThis patch

2007-05-24 Thread Andrew Nagy
That's the problem, as I suspected.  I was using a fresh copy of the 1.1.0.  I 
want to use the MLT feature in my production environment, but I guess I will 
have to roll out the trunk version to my production server.

Thanks!

From: Brian Whitman [EMAIL PROTECTED]
Sent: Thursday, May 24, 2007 10:36 AM
To: solr-user@lucene.apache.org
Subject: Re: compile error with SOLR 69 MoreLikeThis patch

On May 24, 2007, at 10:29 AM, Andrew Nagy wrote:

 That did the trick.  However, now I am trying to apply the patch to
 a fresh copy of solr on another server.  Whenever I do so, I get
 the following error from the patch command:


I just patched trunk two days ago with the latest (posted three days
ago) SOLR-69 patch, cleanly. Are you sure you are trying to patch the
trunk version?




RE: compile error with SOLR 69 MoreLikeThis patch

2007-05-24 Thread Andrew Nagy
Chris, thanks for the tip.  I think I am okay with pushing the trunk to my 
production server.  As we say around here, if you want to be on the bleeding 
edge, you have to be okay with bleeding every once in a while :)

Thanks again
Andrew

From: Chris Hostetter [EMAIL PROTECTED]
Sent: Thursday, May 24, 2007 2:18 PM
To: solr-user@lucene.apache.org
Subject: RE: compile error with SOLR 69 MoreLikeThis patch

: That's the problem, as I suspected.  I was using a fresh copy of the
: 1.1.0.  I want to use the MLT feature in my production environment, but
: I guess I will have to roll out the trunk version to my production
: server.

if your goal is to use the MoreLikeThisHanler then you don't need the
entire trun .. it should work fine with Solr 1.1.0 ... the
mlt functionality added to StandardRequestHandler might be a bit more
challenging, but it should still be possible to retrofit the newer
StandardRequestHandler into the 1.1.0 release if you wnated to.

(Not that i dont' think the trunk is stable, i'm just saying there are
ways to get it working without completley upgrading)



-Hoss



RE: compile error with SOLR 69 MoreLikeThis patch

2007-05-24 Thread Andrew Nagy
duh++

Also, Where should I put the jar file for the patch?

From: Erik Hatcher [EMAIL PROTECTED]
Sent: Thursday, May 24, 2007 2:37 PM
To: solr-user@lucene.apache.org
Subject: Re: compile error with SOLR 69 MoreLikeThis patch

Andrew,

Nightlies are available here: http://people.apache.org/builds/lucene/
solr/nightly/ (a link exists on the wiki main page, for future
reference).

Erik


On May 24, 2007, at 2:28 PM, Andrew Nagy wrote:

 While I am on this topic, I think it might be nice to have a
 nightly build for downloading or is their something like that in
 place, now?
 
 From: Chris Hostetter [EMAIL PROTECTED]
 Sent: Thursday, May 24, 2007 2:18 PM
 To: solr-user@lucene.apache.org
 Subject: RE: compile error with SOLR 69 MoreLikeThis patch

 : That's the problem, as I suspected.  I was using a fresh copy of the
 : 1.1.0.  I want to use the MLT feature in my production
 environment, but
 : I guess I will have to roll out the trunk version to my production
 : server.

 if your goal is to use the MoreLikeThisHanler then you don't need the
 entire trun .. it should work fine with Solr 1.1.0 ... the
 mlt functionality added to StandardRequestHandler might be a bit more
 challenging, but it should still be possible to retrofit the newer
 StandardRequestHandler into the 1.1.0 release if you wnated to.

 (Not that i dont' think the trunk is stable, i'm just saying there are
 ways to get it working without completley upgrading)



 -Hoss



compile error with SOLR 69 MoreLikeThis patch

2007-05-16 Thread Andrew Nagy
I downloaded and patched my solr source with the latest solr69 patch and 
whenever I run ant I get an error:


[javac] 
/office/src/apache-solr-nightly/src/java/org/apache/solr/handler/MoreLikeThisHandler.java:145: 
cannot find symbol

[javac] symbol  : variable DEFAULT_MIN_DOC_FREQ
[javac] location: class org.apache.lucene.search.similar.MoreLikeThis
[javac]   mlt.setMinDocFreq(
params.getInt(MoreLikeThisParams.MIN_DOC_FREQ,  
MoreLikeThis.DEFAULT_MIN_DOC_FREQ));
[javac] 
^



Did I miss something, or maybe I have a non-compatible version of SOLR?

Thanks
Andrew


Ant build question

2007-04-18 Thread Andrew Nagy
Hello, I would like to play with patch SOLR-69 and am trying to rebuild 
solr using ant with some difficulties.  When I try to run ant, I get an 
error saying that it can't find junit, but junit is in the ant lib 
directory.  I had a file called ant-junit.jar in the lib directory and 
copied it to junit.jar, but it still gave me the error.  Any ideas as to 
what is causing this?


Thanks
Andrew

# ant example
Buildfile: build.xml

checkJunitPresence:

BUILD FAILED
/office/src/apache-solr-nightly/build.xml:240: 
##

 JUnit not found.
 Please make sure junit.jar is in ANT_HOME/lib, or made available
 to Ant using other mechanisms like -lib or CLASSPATH.
 ##

Total time: 0 seconds

# echo $ANT_HOME
/usr/local/ant

# ls /usr/local/ant/lib/
ant-antlr.jar ant-apache-oro.jar   ant-commons-net.jar  
ant-javamail.jar  ant-junit.jar ant-starteam.jar   ant-vaj.jar   
junit.jar
ant-apache-bcel.jar   ant-apache-regexp.jarant-icontract.jar
ant-jdepend.jar   ant-launcher.jar  ant-stylebook.jar  ant-weblogic.jar  
README
ant-apache-bsf.jarant-apache-resolver.jar  ant-jai.jar  
ant-jmf.jar   ant-netrexx.jar   ant-swing.jar  ant-xalan1.jar
xercesImpl.jar
ant-apache-log4j.jar  ant-commons-logging.jar  ant.jar  
ant-jsch.jar  ant-nodeps.jarant-trax.jar   ant-xslp.jar  
xml-apis.jar




Re: Ant build question

2007-04-18 Thread Andrew Nagy

Erik Hatcher wrote:

ant-junit.jar != junit.jar

rename it back, and grab junit.jar from junit.org

Easy enough.

Thanks!
Andrew


Re: Schema validator/debugger

2007-04-13 Thread Andrew Nagy

Yonik Seeley wrote:

Oh wait... Andrew, were you always testing via ping?

Check out what the ping query is configured as in solrconfig.xml:

   pingQuery
qt=dismaxamp;q=solramp;start=3amp;fq=id:[* TO *]amp;fq=cat:[* 
TO *]

   /pingQuery

Perhaps we should change it to something simple by default???  q=solr?

That solves the Jetty failure mystery... so it looks like you either
have a tomcat setup problem, or a Solr bug that only shows under
tomcat.


Yes, this is the problem!  Good catch :)  I have been testing via ping.

However this still does not solve my original problem ... I will dig a 
bit more and see what I can find.


Thanks
Andrew


Re: Schema validator/debugger

2007-04-12 Thread Andrew Nagy

Yonik Seeley wrote:

I dropped your schema.xml directly into the Solr example (using
Jetty), fired it up, and everything works fine!?

Okay, I switched over to Jetty and now I get a different error:
SEVERE: org.apache.solr.core.SolrException: undefined field text


Are you sure you are using the stock solrconfig.xml, etc?

Yep.


I apologize ahead of time if I am being blatantly stupid about this. :)


Re: Schema validator/debugger

2007-04-12 Thread Andrew Nagy

Ryan McKinley wrote:

With a clean checkout, you can run:
$ ant example
$ cd example
$ java -jar start.jar

and things work OK.

But, when you delete all but the two fields, you get an exception 
somewhere?
Well, I was working from my own directory, not the example directory.  I 
can give that a try ... but all I did was copy everything from the 
example dir to a different dir.


But, basically ... yes.

I will delete all instances of solr on my server and start from scratch 
... i must have something screwy somewhere.


Re: Schema validator/debugger

2007-04-11 Thread Andrew Nagy

Andrew Nagy wrote:

Ryan McKinley wrote:
What errors are you getting?  Are there exceptions in the log when it 
starts up?

Just a null pointer exception.

I added a few fields to my schema, and then replaced my solr war file 
with the latest build (see my message from a week ago).  It wouldn't 
work, so I assumed something wrong with the latest build.  I then 
reverted back to my 1.1.0 copy and am getting these errors.  The XML 
is valid, but that is all I can test.


I just installed SOLR-75 patch and the schema browser is able to view 
the schema perfectly.  When I used the default schema with SOLR, 
everything is fine, but when I replace my schema ... it's throws this 
NullPointerException 


Any ideas?

Andrew


Re: Schema validator/debugger

2007-04-11 Thread Andrew Nagy

Ryan McKinley wrote:

Are you using the example solrconfig.xml?  The stack trace looks like
an error finding the solr.home index directory, that is configured in
solrconfig.xml, not schema.xml
Yeah, I noticed that too ... but I don't understand why it can't find 
the home.  I have the data home set to ./solr/data and I always start 
tomcat from the parent directory.


If the example schema.xml works, can you try adding a little bit of
your schema at a time?

Yeah, that is my last resort.  I guess I have no choice!

Thanks
Andrew


Re: Schema validator/debugger

2007-04-11 Thread Andrew Nagy

Greg Ludington wrote:

I just installed SOLR-75 patch and the schema browser is able to view
the schema perfectly.  When I used the default schema with SOLR,
everything is fine, but when I replace my schema ... it's throws this
NullPointerException 


One thing to note is that the schema browser as currently in SOLR-75
just transforms the xml file -- it does not represent any of the
actual parsing of the configuration by Solr upon startup, and so would
not be able to catch any errors Solr might encounter in the schema.
Sure, but it's an easier way for my non-textual brain to grasp the 
schema file :)


Nice job, BTW.

Off topic a bit, Has anyone set forth to build a new admin interface for 
SOLR?  I build a lot of admin interfaces for my day job and would love 
to give the admin module a bit of a tune-up (I won't use the term overhaul).





Re: Schema validator/debugger

2007-04-11 Thread Andrew Nagy

Andrew Nagy wrote:

Ryan McKinley wrote:

If the example schema.xml works, can you try adding a little bit of
your schema at a time?

Yeah, that is my last resort.  I guess I have no choice!
So I did this, and I ended up removing all of my fields and copyfields 
with still no luck.  I took a fresh copy of the example schema, replaced 
my fields and copyfields and it still the error persists.


One thing I noticed is that in the admin it goes from Solr Admin 
(example) to Solr Admin (unknown) which makes me think it is having a 
hard time parsing the schema file?


Would it help if I posted my schema file?

Thanks!
Andrew


Re: Schema validator/debugger

2007-04-11 Thread Andrew Nagy

Yonik Seeley wrote:

On 4/11/07, Andrew Nagy [EMAIL PROTECTED] wrote:

 If the example schema.xml works, can you try adding a little bit of
 your schema at a time?
Yeah, that is my last resort.  I guess I have no choice!


That certainly is strange... Sounds like you definitely hit some sort
of Solr bug, but I don't see how an error in schema parsing can be
lost w/o it going into a log somewhere.  Did you check the other
Tomcat log files?

The only tomcat log file that has any mention of an error is catalina.out


Perhaps try putting your schema.xml into the example distributed with
Solr (jetty based) and see what happens.
I can try this.  Will jetty create a log file somewhere? 


Thanks
Andrew


Re: Schema validator/debugger

2007-04-11 Thread Andrew Nagy

Ryan McKinley wrote:


Off topic a bit, Has anyone set forth to build a new admin interface for
SOLR?  I build a lot of admin interfaces for my day job and would love
to give the admin module a bit of a tune-up (I won't use the term 
overhaul).




i think we definitely need an updated admin interface, yes!

Ideally, we could get rid of all the JSP and do everything with
requestHandlers + XSLT.  Then everything could work with only a JRE -
it would not require a JDK and all the jsp jars.  AND this would let
other clients have access to the same info with JSON/xml etc

check:
http://issues.apache.org/jira/browse/SOLR-162

This puts a lot of the functionality as handlers, but still needs an
XSLT front end
Yes, Erik Hatcher showed this off a bit at a conference I was at in 
February.  I was thinking about building off of these xml files.  I do 
quite a bit with XSLT in my day job, so I would love to play around with 
it.  I think this is something that is desperately needed in SOLR, I 
know it would make my life a lot easier :)


I will take a stab at some of this soon ...

Andrew




Schema validator/debugger

2007-04-10 Thread Andrew Nagy

Does anyone have a good method of debugging a schema?
I have been struggling to get my new schema to run for the past couple 
of days and just do not see anything wrong with the XML document.


Thanks
Andrew


SEVERE: Error filterStart

2007-04-05 Thread Andrew Nagy
Hello, I downloaded the latest nightly snapshot of Solr and replaced my 
existing war with the new one.  Once I restarted tomcat, I get this error:


SEVERE: Error filterStart
Apr 5, 2007 10:11:28 AM org.apache.catalina.core.StandardContext start
SEVERE: Context [/solr] startup failed due to previous errors

Any ideas as to what is causing this?  I deleted my index to start with 
a clean slate but I did not change any of my config files, do I need to 
update these or are the backwards compatible?


Thanks!
Andrew



Facet fields question

2007-03-29 Thread Andrew Nagy

Hello

I have 2 fields that I am faceting on, both of which are of type 
string.  The first field is a copyfield from a text field copied to 
a string field for faceting.  The other is purely a string field.  
The faceted results of the copyfield are accurate; however the facet 
results of the 2nd string field all have the ends of the words cut off.  
How can I get the 2nd string field to display in it's original form?  Do 
I need to make a copyField for all fields that are being faceted on?


Also, both fields are multivalued.

Thanks!
Andrew


fetch only facets

2007-03-16 Thread Andrew Nagy
Is their a way to not return any docs and only facets?  I tried setting 
the fl equal to blank, but then i get everything back. 


Thanks
Andrew


cache sizes

2007-03-16 Thread Andrew Nagy
Is their a science to choosing a cache sizes?  I have about 500,000 
records and am seeing a lot of evictions, about 50% of lookups.  What 
factors can i look at to determine what my cache sizes should be?


Here are my cache statistics:

filterCache 
class: org.apache.solr.search.LRUCache 
version: 1.0 
description: LRU Cache(maxSize=10, initialSize=5, 
autowarmCount=5, 
[EMAIL PROTECTED]) 
stats: lookups : 3195074

hits : 1541249
hitratio : 0.48
inserts : 1653824
evictions : 1553824
size : 10
cumulative_lookups : 3195074
cumulative_hits : 1541249
cumulative_hitratio : 0.48
cumulative_inserts : 1653825
cumulative_evictions : 1553825
  
name: queryResultCache 
class: org.apache.solr.search.LRUCache 
version: 1.0 
description: LRU Cache(maxSize=10, initialSize=5000, 
autowarmCount=5000, 
[EMAIL PROTECTED]) 
stats: lookups : 1976

hits : 872
hitratio : 0.44
inserts : 1199
evictions : 0
size : 1104
cumulative_lookups : 1976
cumulative_hits : 872
cumulative_hitratio : 0.44
cumulative_inserts : 1199
cumulative_evictions : 0
  
name: documentCache 
class: org.apache.solr.search.LRUCache 
version: 1.0 
description: LRU Cache(maxSize=512, initialSize=512) 
stats: lookups : 177530

hits : 13255
hitratio : 0.07
inserts : 164275
evictions : 163763
size : 512
cumulative_lookups : 177530
cumulative_hits : 13255
cumulative_hitratio : 0.07
cumulative_inserts : 164275
cumulative_evictions : 163763


Re: Hierarchical Facets

2007-03-07 Thread Andrew Nagy

Hmm ... I had a brain storm.

Could I do something like this:
collectionDir1/Subdir1/SubSubDir1/collection

Then query collection:Dir1/Subdir1 and get the facets on collection at 
that point to see all of the subsubdirectories?


Is their any better method?

Andrew

Andrew Nagy wrote:
I am running into a stumbling block and can only find a way to solve 
the problem with some sort of hierarchical faceting system.  I am in 
the process of moving my records from eXist (http://exist.sf.net) to 
Solr, but the problem is with the lack of a directory structure that 
exist has.  I figure I could easily solve this problem by setting a 
field called collection and then faceting on that field to get my 
list of directories however this will not allow me to do 
subdirectories.


Any ideas on how to implement such a thing in solr?

Thanks!
Andrew



Hierarchical Facets

2007-03-07 Thread Andrew Nagy
I am running into a stumbling block and can only find a way to solve the 
problem with some sort of hierarchical faceting system.  I am in the 
process of moving my records from eXist (http://exist.sf.net) to Solr, 
but the problem is with the lack of a directory structure that exist 
has.  I figure I could easily solve this problem by setting a field 
called collection and then faceting on that field to get my list of 
directories however this will not allow me to do subdirectories.


Any ideas on how to implement such a thing in solr?

Thanks!
Andrew


Installation woes

2007-03-07 Thread Andrew Nagy
Hello, I am trying to install another copy of solr on a server.  I have 
done this many times before, but am running into errors now and I am not 
sure what is causing them.


I unzipped a copy of 1.1.0 and placed the .war file into tomcat.  Then I 
created the solr directory with my bin, data, conf directories.
When I restart tomcat (from the solr directory) and go to the Solr admin 
website everything looks fine.  But, when I pull up the statistics page, 
I only have information in the Query Handlers box, the rest are empty.  
When I click on ping, I see a 500 error from tomcat.


What am I forgetting to do?

Thanks
Andrew


Re: Installation woes

2007-03-07 Thread Andrew Nagy
Argh!  Thanks Yonik for pointing out the log files, duh!  I had a 
malformed line in my schema.xml.  Nice feature to add down the line, 
although I know there is a lot of work going into the admin interface so 
who knows if it is already thought of.  Schema Debugger?  Maybe one day 
I will dig into the code guts and try to contribute.


Andrew


Yonik Seeley wrote:

On 3/7/07, Andrew Nagy [EMAIL PROTECTED] wrote:

Hello, I am trying to install another copy of solr on a server.  I have
done this many times before, but am running into errors now and I am not
sure what is causing them.

I unzipped a copy of 1.1.0 and placed the .war file into tomcat.  Then I
created the solr directory with my bin, data, conf directories.
When I restart tomcat (from the solr directory) and go to the Solr admin
website everything looks fine.  But, when I pull up the statistics page,
I only have information in the Query Handlers box, the rest are empty.
When I click on ping, I see a 500 error from tomcat.

What am I forgetting to do?


Did you point solr to the solr_home (containing your bin, data, and
conf directories)?
Or do you rely on the default of $CWD/solr?

If that's not it, check all the tomcat log files for the first
exception you see.

-Yonik


Re: MoreLikeThis similarity-type queries in Solr

2007-01-31 Thread Andrew Nagy

Bertrand Delacretaz wrote:

On 1/31/07, Brian Whitman [EMAIL PROTECTED] wrote:

Does Solr have support for the Lucene query-contrib MoreLikeThis
query type or anything like it? ...


Yes, there's a patch in http://issues.apache.org/jira/browse/SOLR-69 -
if you try it, please add your comments on that page.


Anyword on something like this being incorporated into the official SOLR 
release?


Andrew


Re: relevance ranking and scoring

2007-01-25 Thread Andrew Nagy

Yonik Seeley wrote:


What about term ranking, could I rank the phrases searched in title
higher than title2?


Absolutely... standard lucene syntax for boosting will give you that
in the standard query handler.

title:(gone with the wind)^3.0 OR title2:(gone with the wind)

That did it!  Thanks for the Help!
What value do the numbers carry in the ranking?  I arbitrarily choose 
the number 5 cause it's an easy number :)


I am a bit nervous about the dismax query system as I have quite a bit 
of other content that could skew the results.
Whats the difference between the dismax query handler and listing all of 
the fields in my search and separating them with an OR?


Thanks!
Andrew




Re: relevance ranking and scoring

2007-01-24 Thread Andrew Nagy

Yonik Seeley wrote:

On 1/23/07, Andrew Nagy [EMAIL PROTECTED] wrote:

Yonik Seeley wrote:
 Things you can try:
 - post the debugging output (including score explain) for the query
I have attached the output.
 - try disabling length normalization for the title field, then remove
 the entire index and re-idnex.
 - try the dismax handler, which can generate sloppy phrase queries to
 boost results containing all terms.
 - try a different similarity implementation
 (org.apache.lucene.misc.SweetSpotSimilarity from lucene)
Can you explain what these 3 options mean?  I would like to get a better
understanding of the guts of SOLR/Lucene but I am too busy working on my
application that uses it to spend time with the internals.


Let's start with the first... add a debugQuery=on
parameter to your request and post the full result here.
You can get the same effect through the
query form on the solr admin pages by checking the Debug: explain 
checkbox.

I attached the results to my last email, are you not able to see them?

Andrew


Re: relevance ranking and scoring

2007-01-24 Thread Andrew Nagy

Yonik Seeley wrote:

Ok, here is your query:
str name=rawquerystringtitle:(gone with the wind) OR title2:(gone
with the wind)/str
And here it is parsed:
str name=parsedquery(title:gone title:wind) (title2:gone 
title2:wind)/str


First, notice how stopwords were removed, so with and the will not
count in the results.

You are querying across two different fields.
Notice how the first two documents both have wind in both title and 
title2,

while the third document gone with the wind has no title2 field (and
hence can't match on it).

In the first documents, the scores for the matches on title and title2
both contribute to the score.  For the third document, it's penalized
by not matching in both the title and title2 field.

You could look at the dismax handler... it helps constructs queries, a
component of which are DisjunctionMaxQueries (they don't add together
scores from different fields, but just take the highest score from any
matching field for a term).

You could also see how changing or removing the stopword list affects 
relevance.

Wow, thanks for the verbose response.  This gives me a lot to go on!

What about term ranking, could I rank the phrases searched in title 
higher than title2?


Thanks!
Andrew


relevance ranking and scoring

2007-01-23 Thread Andrew Nagy

I have 2 questions about the SOLR relevancy system.

1. Why is it when I search for an exact phrase of a title of a record I 
have it generally does not come up as the 1st record in the results?


ex: title:(gone with the wind), the record comes up 3rd.  A record with 
the term wind as the first word in the title comes up 1st.

ex: title:gone with the wind, the record comes up 1st.

Is this because the word wind is the only noun?

2. The score that is associated with each value is quite odd, what 
does it represent.  I generally get results with the top record being 
somewhere around 3.0 or 2.0 and most records are below 1.



Thanks!
Andrew




Re: relevance ranking and scoring

2007-01-23 Thread Andrew Nagy

Yonik Seeley wrote:

Things you can try:
- post the debugging output (including score explain) for the query

I have attached the output.

- try disabling length normalization for the title field, then remove
the entire index and re-idnex.
- try the dismax handler, which can generate sloppy phrase queries to
boost results containing all terms.
- try a different similarity implementation
(org.apache.lucene.misc.SweetSpotSimilarity from lucene)
Can you explain what these 3 options mean?  I would like to get a better 
understanding of the guts of SOLR/Lucene but I am too busy working on my 
application that uses it to spend time with the internals.


Thanks
Andrew
?xml version=1.0 encoding=UTF-8?
response

lst name=responseHeader
 int name=status0/int
 int name=QTime102/int
/lst
result name=response numFound=324 start=0 maxScore=2.7548285
 doc
  float name=score2.7548285/float
  arr name=authorstrFarnol, Jeffery,/str/arr
  str name=callnumberPR6011.A75.W56 1939/str
  str name=formatBook/str
  str name=id97525/str
  str name=languageeng/str
  str name=publishDate1939, c1934./str
  str name=publisherTriangle Books,/str
  str name=titleWinds of chance //str
  str name=title2Winds of change [sic]/str
 /doc
 doc
  float name=score2.5437002/float
  arr name=authorstrSimpson, John E.,/str/arr
  str name=callnumberQC939.L37S56 1994/str
  str name=formatBook/str
  str name=id433518/str
  str name=isbn0521452112/str
  str name=languageeng/str
  str name=physicalill., maps ;/str
  str name=publishDate1994./str
  str name=publisherCambridge University Press,/str
  arr name=subject4astrSea breeze./str/arr
  arr name=subject4xstr//arr
  str name=titleSea breeze and local winds //str
  str name=title2Sea breeze and local wind./str
 /doc
 doc
  float name=score2.438136/float
  arr name=authorstrHobbs, William Herbert,/str/arr
  str name=callnumberG743.H6 1968/str
  str name=formatBook/str
  str name=id192408/str
  str name=languageeng/str
  str name=physicalillus., maps, ports. ;/str
  str name=publishDate[1968, c1930]/str
  str name=publisherGreenwood Press,/str
  arr name=subject4astrMeteorology/str/arr
  arr name=subject4xstr//arr
  arr name=subject5strArctic regions./str/arr
  str name=titleExploring about the North Pole of the winds./str
  str name=title2North Pole of the winds./str
 /doc
 doc
  float name=score2.4319565/float
  arr name=authorstrMitchell, Margaret,/str/arr
  str name=callnumberPS3525.I972G6 1996/str
  str name=formatBook/str
  str name=id426657/str
  str name=isbn0684826259 (alk. paper)/str
  str name=languageeng/str
  str name=physicalill. ;/str
  str name=publishDatec1996./str
  str name=publisherScribner,/str
  arr name=subject4astrWomen/str/arr
  arr name=subject4xstrHistory/str/arr
  arr name=subject5strGeorgia/str/arr
  str name=titleGone with the wind //str
 /doc
 doc
  float name=score2.4319565/float
  arr name=authorstrGable, Clark,/strstrLeigh, Vivien,/strstrHoward, Leslie,/str
	strDe Havilland, Olivia./strstrMitchell, Thomas,/strstrMcDaniel, Hattie,/strstrMcQueen, Butterfly./str
	strFleming, Victor,/strstrMitchell, Margaret,/str/arr
  str name=callnumberVT3188 VHS/str
  str name=formatVideo/str
  str name=id529954/str
  str name=languageeng/str
  str name=physicalsd., col. ;/str
  str name=publishDatec1999./str
  str name=publisherTime Warner Co.,/str
  arr name=subject4astrWar films./strstrFeature films./str/arr
  arr name=subject4xstr/str//arr
  arr name=subject5strUnited States/str/arr
  str name=titleGone with the wind/str
 /doc
 doc
  float name=score2.4319565/float
  arr name=authorstrMitchell, Margaret,/str/arr
  str name=callnumberPS3525.I972G6 1993/str
  str name=formatBook/str
  str name=id534773/str
  str name=isbn0446365386/str
  str name=languageeng/str
  str name=publishDate[1993], c1936./str
  str name=publisherWarner Books,/str
  arr name=subject5strUnited States/strstrGeorgia/str/arr
  str name=titleGone with the wind //str
 /doc
 doc
  float name=score1.7023697/float
  arr name=authorstrPyron, Darden Asbury./str/arr
  str name=callnumberPS3525.I972G687 1983/str
  str name=formatBook/str
  str name=id27783/str
  str name=isbn081300747X (pbk. : alk. paper)/str
  str name=languageeng/str
  str name=publishDatec1983./str
  str name=publisherUniversity Presses of Florida,/str
  arr name=subject1strMitchell, Margaret,/str/arr
  arr name=subject3strGone with the wind (Motion picture)/str/arr
  arr name=subject5strSouthern States/str/arr
  str name=titleRecasting :Gone with the wind in American culture //str
 /doc
 doc
  float name=score1.6493776/float
  arr name=authorstrStuttgarter Bläserquintett./strstrHaydn, Joseph,/strstrReicha, Anton,/str
	strDanzi, Franz,/strstrLickl, Johann Georg,/str/arr
  str name=callnumberCD257/str
  str name=contentsDivertimento, Nr. 1, B-Dur : Chorale St. Antoni / Joseph Haydn (10:24) -- Bläserquintett Es-Dur, op. 88, 2 / Anton Reicha (14:06) -- Bläserquintett B-Dur, op. 56, 1 / Franz Danzi (13:31) -- Quintetto concertante, F-Dur / 

New SOLR installation problems

2006-12-11 Thread Andrew Nagy
I installed the 12-8 snapshot of solr on my 64bit RH AS server and 
whenever I go to the admin page I get the following error:


SEVERE: Servlet.service() for servlet jsp threw exception
java.lang.NoClassDefFoundError: Could not initialize class 
org.apache.solr.core.SolrCore


Any ideas as to what is causing this?

Thanks
Andrew


Re: New SOLR installation problems

2006-12-11 Thread Andrew Nagy
Thanks Yonik for the reply. 
I am using tomcat, and there is nothing in the catalina.out file.  The 
access log just reports the same error I see in the browser which is 
reported below.
I am starting tomcat from my solr directory which has the data, bin and 
conf directories as subdirectories.  So the CWD should be correct.  I 
have this same setup on another server that I have been working on with 
no problem.  Im kinda lost with this one.

Is their a setting in the solrconfig.xml file that I should be looking at?

Andrew

Yonik Seeley wrote:


On 12/11/06, Andrew Nagy [EMAIL PROTECTED] wrote:


I installed the 12-8 snapshot of solr on my 64bit RH AS server and
whenever I go to the admin page I get the following error:

SEVERE: Servlet.service() for servlet jsp threw exception
java.lang.NoClassDefFoundError: Could not initialize class
org.apache.solr.core.SolrCore

Any ideas as to what is causing this?



Look through the logs of whatever servlet container you are using for
the first exception thrown.  It most likely has something to do with
not being able to find the solr config files.

-Yonik




Re: New SOLR installation problems

2006-12-11 Thread Andrew Nagy
Nevermind, I got it working now.  Had the paths setup incorrectly. 
Dumb++


Andrew

Andrew Nagy wrote:

Thanks Yonik for the reply. I am using tomcat, and there is nothing in 
the catalina.out file.  The access log just reports the same error I 
see in the browser which is reported below.
I am starting tomcat from my solr directory which has the data, bin 
and conf directories as subdirectories.  So the CWD should be 
correct.  I have this same setup on another server that I have been 
working on with no problem.  Im kinda lost with this one.
Is their a setting in the solrconfig.xml file that I should be looking 
at?


Andrew

Yonik Seeley wrote:


On 12/11/06, Andrew Nagy [EMAIL PROTECTED] wrote:


I installed the 12-8 snapshot of solr on my 64bit RH AS server and
whenever I go to the admin page I get the following error:

SEVERE: Servlet.service() for servlet jsp threw exception
java.lang.NoClassDefFoundError: Could not initialize class
org.apache.solr.core.SolrCore

Any ideas as to what is causing this?




Look through the logs of whatever servlet container you are using for
the first exception thrown.  It most likely has something to do with
not being able to find the solr config files.

-Yonik





multiple collections

2006-12-11 Thread Andrew Nagy
I was wondering how I might create multiple collections that have 
different field sets under solr.  Would I have to have multiple 
implementations of solr running, or can I have more than one schema.xml 
file per collection ?


Thanks
Andrew


Re: Result: numFound inaccuracies

2006-12-09 Thread Andrew Nagy
- Original Message -
From: Yonik Seeley [EMAIL PROTECTED]
Date: Friday, December 8, 2006 6:01 pm
Subject: Re: Result: numFound inaccuracies
To: solr-user@lucene.apache.org

 
 start is 0 based :-)
 

Man do I feel dumb!

Andrew


Re: Facet Performance

2006-12-08 Thread Andrew Nagy

Yonik Seeley wrote:


1) facet on single-valued strings if you can
2) if you can't do (1) then enlarge the fieldcache so that the number
of filters (one per possible term in the field you are filtering on)
can fit.


I changed the filterCache to the following:
   filterCache
 class=solr.LRUCache
 size=25600
 initialSize=5120
 autowarmCount=1024/

However a search that normally takes .04s is taking 74 seconds once I 
use the facets since I am faceting on 4 fields.


Can you suggest a better configuration that would solve this performance 
issue, or should I not use faceting?
I figure I could run the query twice, once limited to 20 records and 
then again with the limit set to the total number of records and develop 
my own facets.  I have infact done this before with a different back-end 
and my code is processed in under .01 seconds.


Why is faceting so slow?

Andrew


Re: Facet Performance

2006-12-08 Thread Andrew Nagy

Chris Hostetter wrote:


: Could you suggest a better configuration based on this?

If that's what your stats look like after a single request, then i would
guess you would need to make your cache size at least 1.6 million in order
for it to be of any use in improving your facet speed.
 

Would this have any strong impacts on my system?  Should I just set it 
to an even 2 million to allow for growth?



: My data is 492,000 records of book data.  I am faceting on 4 fields:
: author, subject, language, format.
: Format and language are fairly simple as their are only a few unique
: terms.  Author and subject however are much different in that there are
: thousands of unique terms.

by the looks of it, you have a lot more then a few thousand unique terms
in those two fields ... are you tokenizing on these fields?  that's
probably not what you want for ields you're going to facet on.
 

All of these fields are set as string in my schema, so if I understand 
the fields correctly, they are not being tokenized.  I also have an 
author field that is set as text for searching.


Thanks
Andrew


Re: Facet Performance

2006-12-08 Thread Andrew Nagy

Yonik Seeley wrote:


Are they multivalued, and do they need to be.
Anything that is of type string and not multivalued will use the
lucene FieldCache rather than the filterCache.


The author field is multivalued.  Will this be a strong performance issue?

I could make multiple author fields as to not have the multivalued field 
and then only facet on the first author.


Thanks
Andrew




Re: Facet Performance

2006-12-08 Thread Andrew Nagy

J.J. Larrea wrote:


Unfortunately which strategy will be chosen is currently undocumented and 
control is a bit oblique:  If the field is tokenized or multivalued or Boolean, 
the FilterQuery method will be used; otherwise the FieldCache method.  I expect 
I or others will improve that shortly.
 

Good to hear, cause I can't really get away with not having a 
multi-valued field for author.


Im really excited by solr and really impressed so far.

Thanks!
Andrew


Re: Facet Performance

2006-12-08 Thread Andrew Nagy

Erik Hatcher wrote:


On Dec 8, 2006, at 2:15 PM, Andrew Nagy wrote:

My data is 492,000 records of book data.  I am faceting on 4  fields: 
author, subject, language, format.
Format and language are fairly simple as their are only a few  unique 
terms.  Author and subject however are much different in  that there 
are thousands of unique terms.



When encountering difficult issues, I like to think in terms of the  
user interface.  Surely you're not presenting 400k+ authors to the  
users in one shot.  In Collex, we have put an AJAX drop-down that  
shows the author facet (we call it name on the UI, with various roles  
like author, painter, etc).  You can see this in action here:


In our data, we don't have unique authors for each records ... so let's 
say out of the 500,000 records ... we have 200,000 authors.  What I am 
trying to display is the top 10 authors from the results of a search.  
So I do a search for title:Gone with the wind and I would like to see 
the top 10 matching authors from these results.


But no worries, I have written my own facet handler and I am now back to 
under a second with faceting!


Thanks for everyone's help and keep up the good work!

Andrew


Result: numFound inaccuracies

2006-12-08 Thread Andrew Nagy

Hello, me again.

I have been running some extensive tests of my search engine and have 
been seeing inaccuracies with the numFound attribute.  It tends to 
return 1 more than what is actually show in the XML.


Is this a bug, or could I be doing something wrong?

I have a specific example in front of me at the moment where my query 
found 2 records, yet I get: result numFound=3 start=1


Any ideas?

Andrew


Facet Performance

2006-12-07 Thread Andrew Nagy
In September there was a thread [1] on this list about heterogeneous 
facets and their performance.  I am having a similar issue and am 
unclear as the resolution of this thread.


I performed a search against my dataset (492,000 records) and got the 
results I am looking for in .3 seconds.  I then set facet to true and 
got results in 16 seconds and the facets include data that is not in my 
result set, it is from the entire set.  How do I limit the faceting to 
my results set and speed up the results?


Thanks!
Andrew

[1] http://www.mail-archive.com/solr-user@lucene.apache.org/msg00955.html


Changing the schema.xml

2006-12-06 Thread Andrew Nagy
What is necessary for the effects of changing the schema.xml to take 
effect for all of my records?  I restarted tomcat, but it does not seem 
that my changes have taken effect.


I wanted to change a full-text field from type:string to type:text to 
allow for better searching, but do no see any differences when I make my 
changes to the schema.xml.


Thanks!
Andrew


Initial import problems

2006-12-05 Thread Andrew Nagy

Hello, I am new to SOLR but very excited for it's possibilities.

I am having some difficulties with my data import which I hope can be 
solved very easily.
First I wrote an xslt to transform my xml into the solr schema and 
modified the schema.xml to match the fields that I created.  I then ran 
the post.sh on my 492,000 records that I have.  Near the end of the 
process the records stopped being added due to a memory heap error.  I 
obviously maxed the allotted memory for the import.  Next time I will 
import less at a time!


I then posted a commit statement.  I went to my solr admin site and 
looked at the statistics.  It said 372,000 records (roughly) were stored 
and 1 commit.  I tried to do a search but no matter what I search for I 
get 0 results.  I even tried title:the (assuming it is not blocking 
the stop word, it should return something!).


It appears to me that the search is not searching any records.  Any idea 
as to what I might need to do, or should I start over from scratch and 
re-import my records in smaller chunks?


Thanks!
Andrew