Re: Question on StreamingUpdateSolrServer

2009-04-11 Thread Shalin Shekhar Mangar
On Sat, Apr 11, 2009 at 3:29 AM, vivek sar vivex...@gmail.com wrote:

 I also noticed that the Solr app has over 6000 file handles open -

lsof | grep solr | wc -l   - shows 6455

 I've 10 cores (using multi-core) managed by the same Solr instance. As
 soon as start up the Tomcat the open file count goes up to 6400.  Few
 questions,

 1) Why is Solr holding on to all the segments from all the cores - is
 it because of auto-warmer?


You have 10 cores, so Solr opens 10 indexes, each of which contains multiple
files. That is one reason. Apart from that, Tomcat will keep some file
handles for incoming connections.



 2) How can I reduce the open file count?


Are they causing a problem? Tomcat will log messages when it cannot accept
incoming connections if it runs out of available file handles. But if you
experiencing issues, you can increase the file handle limit or you can set
useCompoundFile=true in solrconfig.xml.



 3) Is there a way to stop the auto-warmer?
 4) Could this be related to Tomcat returning blank page for every
 request?


It could be. Check the Tomcat and Solr logs.

-- 
Regards,
Shalin Shekhar Mangar.


Re: sorlj search

2009-04-11 Thread Shalin Shekhar Mangar
On Wed, Feb 6, 2008 at 10:51 AM, Tevfik Kiziloren tevfikkizilo...@gmail.com
 wrote:


 Caused by: org.apache.solr.common.SolrException: parsing error
at

 org.apache.solr.client.solrj.impl.XMLResponseParser.processResponse(XMLResponseParser.java:138)
at

 org.apache.solr.client.solrj.impl.XMLResponseParser.processResponse(XMLResponseParser.java:99)
at

 org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.request(CommonsHttpSolrServer.java:317)
at

 org.apache.solr.client.solrj.request.QueryRequest.process(QueryRequest.java:84)
... 29 more
 Caused by: java.lang.RuntimeException: this must be known type! not: int
at

 org.apache.solr.client.solrj.impl.XMLResponseParser.readNamedList(XMLResponseParser.java:217)
at

 org.apache.solr.client.solrj.impl.XMLResponseParser.readNamedList(XMLResponseParser.java:235)
at

 org.apache.solr.client.solrj.impl.XMLResponseParser.processResponse(XMLResponseParser.java:123)



Which version of Solr and Solrj client are you using?

-- 
Regards,
Shalin Shekhar Mangar.


Re: Help with relevance failure in Solr 1.3

2009-04-11 Thread Grant Ingersoll


On Apr 10, 2009, at 5:50 PM, Walter Underwood wrote:


Normally, both changeling and the changeling work fine. This one
server is misbehaving like this for all multi-term queries.

Yes, it is VERY weird that the term changeling does not show up in
the explain.

A server will occasionally go bad and stay in that state. In one  
case,

two servers went bad and both gave the same wrong results.



What's the solution for when they go bad?  Do you have to restart Solr  
or reboot or what?




Here is the dismax config. groups means movies. The title* fields
are stemmed and stopped, the exact* fields are not.

 !-- groups and people  --

 requestHandler name=groups_people class=solr.SearchHandler
   lst name=defaults
str name=defTypedismax/str
str name=echoParamsnone/str
float name=tie0.01/float
str name=qf
   exact^6.0 exact_alt^6.0 exact_base~jw_0.7_1^8.0 exact_alias^8.0
title^3.0 title_alt^3.0 title_base^4.0
/str

str name=pf
   exact^9.0 exact_alt^9.0 exact_base^12.0 exact_alias^12.0  
title^3.0

title_alt^4.0 title_base^6.0
/str
str name=bf
   search_popularity^100.0
/str
str name=mm1/str
int name=ps100/int
str name=flid,type,movieid,personid,genreid/str

   /lst
   lst name=appends
 str name=fqtype:group OR type:person/str
   /lst
 /requestHandler


wunder

On 4/10/09 12:51 PM, Grant Ingersoll gsing...@apache.org wrote:



On Apr 10, 2009, at 1:56 PM, Walter Underwood wrote:


We have a rare, hard-to-reproduce problem with our Solr 1.3 servers,
and
I would appreciate any ideas.

Ocassionally, a server will start returning results with really poor
relevance. Single term queries work fine, but multi-term queries are
scored based on the most common term (lowest IDF).

I don't see anything in the logs when this happens. We have a  
monitor

doing a search for the 100 most popular movies once per minute to
catch this, so we know when it was first detected.

I'm attaching two explain outputs, one for the query changeling  
and

one for the changeling.



I'm not sure what exactly  you are asking, so bear with me...

Are you saying that the changeling normally returns results just
fine and then periodically it will go bad or are you saying you
don't understand why the changeling scores differently from
changeling?  In looking at the explains, it is weird that in the
the changeling case, the term changeling doesn't even show up as a
term.

Can you share your dismax configuration?  That will be easier to  
parse

than trying to make sense of the debug query parsing.

-Grant




--
Grant Ingersoll
http://www.lucidimagination.com/

Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids)  
using Solr/Lucene:

http://www.lucidimagination.com/search



Re: Index Version Number

2009-04-11 Thread Yonik Seeley
On Fri, Apr 10, 2009 at 11:58 AM, Richard Wiseman
rwise...@infosciences.com wrote:
 Is it possible for a Solr client to determine if the index has changed since
 the last time it performed a query?  For example, is it possible to query
 the current Lucene indexVersion?

Grant pointed to one way - the Luke handler.
Another way is to look at the Last-Modified or ETag HTTP headers.

$ curl -i http://localhost:8983/solr/select?q=solr
HTTP/1.1 200 OK
Last-Modified: Fri, 10 Apr 2009 17:40:54 GMT
ETag: OWZlNjdkN2Q4ODAwMDAwU29scg==
Content-Type: text/xml; charset=utf-8
Content-Length: 2308
Server: Jetty(6.1.3)


-Yonik
http://www.lucidimagination.com


Re: Its urgent! plz help in schema.xml- appending one field to another

2009-04-11 Thread Udaya



Erik Hatcher wrote:
 
 
 On Apr 8, 2009, at 9:50 PM, Udaya wrote:
 

 Hi,
 Need your help,
 I would like to know how we could append or add one field value to  
 another
 field in Scheme.xml
 My scheme is as follows (only the field part is given):
 Scheme.xml
 fields
 field name=topics_id type=integer indexed=true stored=true
 required=true /
   field name=topics_subject type=text indexed=true  
 stored=true
 required=true/
   field name=post_text type=text indexed=true stored=true
 multiValued=true/

field name=url type=string stored=true
 default=http://comp.com/portals/ForumWindow? 
 action=1v=tp=topics_id#topics_id
 /
field name=all_text type=text indexed=true stored=true
 multiValued=true/

 Here for the field with name topics_id we get id from a table. I  
 what his
 topics_id value to be appended into the default value attribute of  
 the field
 with name url.

 For eg:
 Suppose if we get topics_id value as 512 during a search then the  
 value of
 the url should be appended as
 http://comp.com/portals/JBossForumWindow?action=1v=tp=512#512

 Is this possible, plz give me some suggestions.
 
 If you're using DIH to index your table, you could aggregate using the  
 template transformer during indexing.
 
 If you're indexing a different way, why not let the searching client  
 (UI) do the aggregation of an id into a URL?
 
   Erik
 
 
 

Hi Erik thanks for your suggestion :-)... i do have appended the id into the
url by writing code in solrj and java...
thanks
Udaya

-- 
View this message in context: 
http://www.nabble.com/Its-urgent%21-plz-help-in-schema.xml--appending-one-field-to-another-tp22963108p23001380.html
Sent from the Solr - User mailing list archive at Nabble.com.



Evaluating Solr

2009-04-11 Thread Chris Masters

Hi All,

I'm evaluating Solr to provide a single fast and effective searching solution 
for a range of web applications (Java, .Net  PHP). So a couple questions:

1) Can we configure Solr to auto reindex/import if we point it to a datasource? 
I've done some searching on this and have only come up with specific mailing 
list questions on DataImporter - there seems to be no docs on that. If so, can 
we use this mechnism to point to multiple datasources for different indexes 
i.e. use datasource A for index A and datasource B for index B? 

2) Can we create multiple indexes of the same objects i.e. index of users for 
customer A and another for users of customer B? If so, how would be specify 
context/index to use?

3) It seems that the built-in replication depends on rsync; we use a SAN for 
our production databases, can we easily use the SAN for replication/redundancy? 
Are there fiel system locking issues (as with out-of-the-box lucene) with two 
Solr servers using the same index shared via a SAN? 

4) We've extended one of our local lucene search implementations to store the 
index in a database (not for speed but for redundancy  load balancing) using a 
JDBCDirectoryProvider implementation - is it possible to use that with Solr?

Thanks for any help on this!

rotis 23



  


Re: Help with relevance failure in Solr 1.3

2009-04-11 Thread Walter Underwood
Restarting Solr fixes it. If I remember correctly, a sync and commit
does not fix it. I have disabled snappuller this time, so I can study
the broken instance.

wunder

On 4/11/09 5:03 AM, Grant Ingersoll gsing...@apache.org wrote:

 
 On Apr 10, 2009, at 5:50 PM, Walter Underwood wrote:
 
 Normally, both changeling and the changeling work fine. This one
 server is misbehaving like this for all multi-term queries.
 
 Yes, it is VERY weird that the term changeling does not show up in
 the explain.
 
 A server will occasionally go bad and stay in that state. In one
 case,
 two servers went bad and both gave the same wrong results.
 
 
 What's the solution for when they go bad?  Do you have to restart Solr
 or reboot or what?
 
 
 Here is the dismax config. groups means movies. The title* fields
 are stemmed and stopped, the exact* fields are not.
 
  !-- groups and people  --
 
  requestHandler name=groups_people class=solr.SearchHandler
lst name=defaults
 str name=defTypedismax/str
 str name=echoParamsnone/str
 float name=tie0.01/float
 str name=qf
exact^6.0 exact_alt^6.0 exact_base~jw_0.7_1^8.0 exact_alias^8.0
 title^3.0 title_alt^3.0 title_base^4.0
 /str
 
 str name=pf
exact^9.0 exact_alt^9.0 exact_base^12.0 exact_alias^12.0
 title^3.0
 title_alt^4.0 title_base^6.0
 /str
 str name=bf
search_popularity^100.0
 /str
 str name=mm1/str
 int name=ps100/int
 str name=flid,type,movieid,personid,genreid/str
 
/lst
lst name=appends
  str name=fqtype:group OR type:person/str
/lst
  /requestHandler
 
 
 wunder
 
 On 4/10/09 12:51 PM, Grant Ingersoll gsing...@apache.org wrote:
 
 
 On Apr 10, 2009, at 1:56 PM, Walter Underwood wrote:
 
 We have a rare, hard-to-reproduce problem with our Solr 1.3 servers,
 and
 I would appreciate any ideas.
 
 Ocassionally, a server will start returning results with really poor
 relevance. Single term queries work fine, but multi-term queries are
 scored based on the most common term (lowest IDF).
 
 I don't see anything in the logs when this happens. We have a
 monitor
 doing a search for the 100 most popular movies once per minute to
 catch this, so we know when it was first detected.
 
 I'm attaching two explain outputs, one for the query changeling
 and
 one for the changeling.
 
 
 I'm not sure what exactly  you are asking, so bear with me...
 
 Are you saying that the changeling normally returns results just
 fine and then periodically it will go bad or are you saying you
 don't understand why the changeling scores differently from
 changeling?  In looking at the explains, it is weird that in the
 the changeling case, the term changeling doesn't even show up as a
 term.
 
 Can you share your dismax configuration?  That will be easier to
 parse
 than trying to make sense of the debug query parsing.
 
 -Grant
 
 
 --
 Grant Ingersoll
 http://www.lucidimagination.com/
 
 Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids)
 using Solr/Lucene:
 http://www.lucidimagination.com/search
 



Re: Evaluating Solr

2009-04-11 Thread Shalin Shekhar Mangar
On Sat, Apr 11, 2009 at 7:26 PM, Chris Masters roti...@yahoo.com wrote:


 I'm evaluating Solr to provide a single fast and effective searching
 solution for a range of web applications (Java, .Net  PHP).


That's good news for Solr :)


 So a couple questions:

 1) Can we configure Solr to auto reindex/import if we point it to a
 datasource? I've done some searching on this and have only come up with
 specific mailing list questions on DataImporter - there seems to be no docs
 on that. If so, can we use this mechnism to point to multiple datasources
 for different indexes i.e. use datasource A for index A and datasource B for
 index B?


There's a comprehensive wiki document at
http://wiki.apache.org/solr/DataImportHandler


 2) Can we create multiple indexes of the same objects i.e. index of users
 for customer A and another for users of customer B? If so, how would be
 specify context/index to use?


Looks like you need http://wiki.apache.org/solr/MultipleIndexes


 3) It seems that the built-in replication depends on rsync; we use a SAN
 for our production databases, can we easily use the SAN for
 replication/redundancy? Are there fiel system locking issues (as with
 out-of-the-box lucene) with two Solr servers using the same index shared via
 a SAN?


The issues would be same as for Lucene. I've never used Solr with SAN so I
do not know about if it will work or how well it will perform.


 4) We've extended one of our local lucene search implementations to store
 the index in a database (not for speed but for redundancy  load balancing)
 using a JDBCDirectoryProvider implementation - is it possible to use that
 with Solr?


Again, never tried. Though, support for configurable directory provider was
added recently (in trunk). See
https://issues.apache.org/jira/browse/SOLR-465 for more details.

-- 
Regards,
Shalin Shekhar Mangar.


DIH XML/update

2009-04-11 Thread Antonio Eggberg

Hi,

Wondering if there is a solution to this. I am using DIH to index an XML feed, 
I update the feed every 10 min and now I have an index of more then 10 million 
docs, I do DIH update/clean = false. Now more and more I update my indexing 
time is increasing and its coming to a point where I can see indexing time will 
be greater then the update time. 

Do you have any suggestion? I am thinking of creating a large index and then 
every X hours merge the small index with the large ones -- small index is the 
one which gets updated by the feed. any other thoughts? How are you folks doing 
it?

Regards
Antonio


  __
Låna pengar utan säkerhet. Jämför vilkor online hos Kelkoo.
http://www.kelkoo.se/c-100390123-lan-utan-sakerhet.html?partnerId=96915014


Evaluating Solr and Ruby-Solr

2009-04-11 Thread Thanh Doan
We are also Evaluating Solr and Sphinx to see which one provides a suitable
search solution for our ROR site.

As a newbie I picked the Advanced Rails Recipes book and follow Recipe 27
before building our prototype.

As I followed Recipe 27 of the Advanced Rails Recipes book. I tried to
follow the steps exactly to see if I can drill down the search results using
facets.

Initially solr and solr-ruby appeared to work fine. The below statements
worked great and I was able to query the solr index from Ruby.

solr = Solr::Connection.new(’http://127.0.0.1:8983/solr’)

response = solr.query(’pragmatic’, :facets = {:fields =
['pubisher_facet'], :mincount =1}) docs = response.hits

But when I ran
response.facet_fields(’publisher_facet’)

I got the below error

response.field_facets(’publisher_facet’) NoMethodError: undefined method
`size’ for nil:NilClass from
C:/Ruby/lib/ruby/gems/1.8/gems/solr-ruby-0.0.7/lib/solr/util.rb:17:in
`paired_array_each’ from
C:/Ruby/lib/ruby/gems/1.8/gems/solr-ruby-0.0.7/lib/solr/response/standard.rb:43:in
`field_facets’ from (irb):13

As development box, we currently use solr 1.3, Ruby 1.8.6 on Windows Vista,
solr-ruby 0.0.7.

Please advise if I missed something here wnen follow the recipe?

Thank you.

Thanh


Below are installed gems on my pc.

C:\Users\thanhruby -v
ruby 1.8.6 (2008-08-11 patchlevel 287) [i386-mswin32]

C:\Users\thanhgem -v
1.3.1

C:\Users\thanhgem list

*** LOCAL GEMS ***

actionmailer (2.3.2)
actionpack (2.3.2)
activerecord (2.3.2)
activeresource (2.3.2)
activesupport (2.3.2)
capistrano (2.5.5)
cgi_multipart_eof_fix (2.5.0)
fxri (0.3.6)
fxruby (1.6.16)
gem_plugin (0.2.3)
highline (1.5.0)
hpricot (0.6.164)
log4r (1.0.5)
mongrel (1.1.5)
mongrel_cluster (1.0.5)
net-scp (1.0.2)
net-sftp (2.0.2)
net-ssh (2.0.11)
net-ssh-gateway (1.0.1)
ptools (1.1.6)
rails (2.3.2)
rake (0.8.4, 0.8.1)
ruby-opengl (0.60.0)
solr-ruby (0.0.7)
test-unit (2.0.1)
win32-api (1.2.1, 1.2.0)
win32-clipboard (0.4.4)
win32-dir (0.3.2)
win32-eventlog (0.5.0)
win32-file (0.5.5)
win32-file-stat (1.3.1)
win32-process (0.5.9)
win32-sapi (0.1.4)
win32-sound (0.4.1)
windows-api (0.2.4)
windows-pr (0.9.3)


Re: Evaluating Solr

2009-04-11 Thread Chris Masters

Thanks for the info Shalin - most useful!

With regard to https://issues.apache.org/jira/browse/SOLR-465; I see it's been 
reopened - roughly when is v1.4 planned for release? 

Thanks, rotis



- Original Message 
From: Shalin Shekhar Mangar shalinman...@gmail.com
To: solr-user@lucene.apache.org
Sent: Saturday, April 11, 2009 7:13:11 PM
Subject: Re: Evaluating Solr

On Sat, Apr 11, 2009 at 7:26 PM, Chris Masters roti...@yahoo.com wrote:


 I'm evaluating Solr to provide a single fast and effective searching
 solution for a range of web applications (Java, .Net  PHP).


That's good news for Solr :)


 So a couple questions:

 1) Can we configure Solr to auto reindex/import if we point it to a
 datasource? I've done some searching on this and have only come up with
 specific mailing list questions on DataImporter - there seems to be no docs
 on that. If so, can we use this mechnism to point to multiple datasources
 for different indexes i.e. use datasource A for index A and datasource B for
 index B?


There's a comprehensive wiki document at
http://wiki.apache.org/solr/DataImportHandler


 2) Can we create multiple indexes of the same objects i.e. index of users
 for customer A and another for users of customer B? If so, how would be
 specify context/index to use?


Looks like you need http://wiki.apache.org/solr/MultipleIndexes


 3) It seems that the built-in replication depends on rsync; we use a SAN
 for our production databases, can we easily use the SAN for
 replication/redundancy? Are there fiel system locking issues (as with
 out-of-the-box lucene) with two Solr servers using the same index shared via
 a SAN?


The issues would be same as for Lucene. I've never used Solr with SAN so I
do not know about if it will work or how well it will perform.


 4) We've extended one of our local lucene search implementations to store
 the index in a database (not for speed but for redundancy  load balancing)
 using a JDBCDirectoryProvider implementation - is it possible to use that
 with Solr?


Again, never tried. Though, support for configurable directory provider was
added recently (in trunk). See
https://issues.apache.org/jira/browse/SOLR-465 for more details.

-- 
Regards,
Shalin Shekhar Mangar.



  


Re: Evaluating Solr and Ruby-Solr

2009-04-11 Thread Erik Hatcher


On Apr 11, 2009, at 3:05 PM, Thanh Doan wrote:

We are also Evaluating Solr and Sphinx to see which one provides a  
suitable

search solution for our ROR site.

As a newbie I picked the Advanced Rails Recipes book and follow  
Recipe 27

before building our prototype.

As I followed Recipe 27 of the Advanced Rails Recipes book. I tried to
follow the steps exactly to see if I can drill down the search  
results using

facets.

Initially solr and solr-ruby appeared to work fine. The below  
statements

worked great and I was able to query the solr index from Ruby.

solr = Solr::Connection.new(’http://127.0.0.1:8983/solr’)

response = solr.query(’pragmatic’, :facets = {:fields =
['pubisher_facet'], :mincount =1}) docs = response.hits

But when I ran
response.facet_fields(’publisher_facet’)

I got the below error

response.field_facets(’publisher_facet’) NoMethodError: undefined  
method

`size’ for nil:NilClass from


It's field_facets, not facet_fields.  Using Solr's example data, solr- 
ruby 0.0.7, and irb -


 Solr::Connection.new.query('ipod', :facets = {:fields =  
['cat']}).field_facets('cat').each {|f| p #{f.name} (#{f.value})}

electronics (3)
connector (2)
music (1)
camera (0)
card (0)
copier (0)
drive (0)
graphics (0)
hard (0)
memory (0)
monitor (0)
multifunction (0)
printer (0)
scanner (0)
search (0)
software (0)

Erik



Re: Question on StreamingUpdateSolrServer

2009-04-11 Thread vivek sar
Thanks Shalin.

The problem is I don't see any error message in the catalina.out. I
don't even see the request coming in - I simply get blank page on
browser. If I keep trying the request goes through and I get respond
from Solr, but then it become unresponsive again or sometimes throws
connection reset error. I'm not sure why would it work sometimes and
not the other times for the same query. As soon as I stop the Indexer
process things start working fine. Any way I can debug this problem?

-vivek

On Fri, Apr 10, 2009 at 11:05 PM, Shalin Shekhar Mangar
shalinman...@gmail.com wrote:
 On Sat, Apr 11, 2009 at 3:29 AM, vivek sar vivex...@gmail.com wrote:

 I also noticed that the Solr app has over 6000 file handles open -

    lsof | grep solr | wc -l   - shows 6455

 I've 10 cores (using multi-core) managed by the same Solr instance. As
 soon as start up the Tomcat the open file count goes up to 6400.  Few
 questions,

 1) Why is Solr holding on to all the segments from all the cores - is
 it because of auto-warmer?


 You have 10 cores, so Solr opens 10 indexes, each of which contains multiple
 files. That is one reason. Apart from that, Tomcat will keep some file
 handles for incoming connections.



 2) How can I reduce the open file count?


 Are they causing a problem? Tomcat will log messages when it cannot accept
 incoming connections if it runs out of available file handles. But if you
 experiencing issues, you can increase the file handle limit or you can set
 useCompoundFile=true in solrconfig.xml.



 3) Is there a way to stop the auto-warmer?
 4) Could this be related to Tomcat returning blank page for every
 request?


 It could be. Check the Tomcat and Solr logs.

 --
 Regards,
 Shalin Shekhar Mangar.



Re: Question on StreamingUpdateSolrServer

2009-04-11 Thread Shalin Shekhar Mangar
On Sun, Apr 12, 2009 at 2:15 AM, vivek sar vivex...@gmail.com wrote:


 The problem is I don't see any error message in the catalina.out. I
 don't even see the request coming in - I simply get blank page on
 browser. If I keep trying the request goes through and I get respond
 from Solr, but then it become unresponsive again or sometimes throws
 connection reset error. I'm not sure why would it work sometimes and
 not the other times for the same query. As soon as I stop the Indexer
 process things start working fine. Any way I can debug this problem?


I'm not sure. I've never seen this issue myself.

Could you try using the bundled jetty instead of Tomcat or on a different
box just to make sure this is not an environment specific issue?

-- 
Regards,
Shalin Shekhar Mangar.


Re: Evaluating Solr

2009-04-11 Thread Shalin Shekhar Mangar
On Sun, Apr 12, 2009 at 1:49 AM, Chris Masters roti...@yahoo.com wrote:


 With regard to https://issues.apache.org/jira/browse/SOLR-465; I see it's
 been reopened


Yes, there are issues related to hard coded FSDirectory usage. There's patch
in there but I haven't looked at it yet.


 - roughly when is v1.4 planned for release?


That's a tough one. It is not something that I control but a few days ago I
had answered mid next month for the same question :)

The right answer is: when all the issues marked for 1.4 are resolved and
there is consensus in the community for a release. You can see the list of
unresolved issues at the following url.

https://issues.apache.org/jira/secure/IssueNavigator.jspa?reset=truemode=hidesorter/order=DESCsorter/field=priorityresolution=-1pid=12310230fixfor=12313351

-- 
Regards,
Shalin Shekhar Mangar.


Re: maxCodeLength in PhoneticFilterFactory

2009-04-11 Thread Yonik Seeley
There's definitely a bug - I just reproduced it.  Nothing obvious
jumps out at me... and there's no error in the logs either (that's
another bug it would seem).  Could you open a JIRA issue for this?


-Yonik
http://www.lucidimagination.com



On Fri, Apr 10, 2009 at 6:54 PM, Brian Whitman br...@echonest.com wrote:
 i have this version of solr running:

 Solr Implementation Version: 1.4-dev 747554M - bwhitman - 2009-02-24
 16:37:49

 and am trying to update a schema to support 8 code length metaphone instead
 of 4 via this (committed) issue:

 https://issues.apache.org/jira/browse/SOLR-813

 So I change the schema to this (knowing that I have to reindex)

        filter class=solr.PhoneticFilterFactory encoder=DoubleMetaphone
 inject=true maxCodeLength=8/

 But when I do queries fail with

 Error_initializing_DoubleMetaphoneclass_orgapachecommonscodeclanguageDoubleMetaphone__at_orgapachesolranalysisPhoneticFilterFactoryinitPhoneticFilterFactoryjava90__at_orgapachesolrschemaIndexSchema$6initIndexSchemajava821__at_orgapachesolrschemaIndexSchema$6initIndexSchemajava817__at_orgapachesolrutilpluginAbstractPluginLoaderloadAbstractPluginLoaderjava149__at_orgapachesolrschemaIndexSchemareadAnalyzerIndexSchemajava831__at_orgapachesolrschemaIndexSchemaaccess$100IndexSchemajava58__at_orgapachesolrschemaIndexSchema$1createIndexSchemajava425__at_orgapachesolrschemaIndexSchema$1createIndexSchemajava410__at_orgapachesolrutilpluginAbstractPluginLoaderloadAbstractPluginLoaderjava141__at_orgapachesolrschemaIndexSchemareadSchemaIndexSchemajava452__at_orgapachesolrschemaIndexSchemainitIndexSchemajava95__at_orgapachesolrcoreSolrCoreinitSolrCorejava501__at_orgapachesolrcoreCoreContainer$InitializerinitializeCoreContainerjava121



Re: maxCodeLength in PhoneticFilterFactory

2009-04-11 Thread Yonik Seeley
OK, should hopefully be fixed in trunk.


-Yonik
http://www.lucidimagination.com


On Sat, Apr 11, 2009 at 9:16 PM, Yonik Seeley yo...@apache.org wrote:
 There's definitely a bug - I just reproduced it.  Nothing obvious
 jumps out at me... and there's no error in the logs either (that's
 another bug it would seem).  Could you open a JIRA issue for this?


 -Yonik
 http://www.lucidimagination.com



 On Fri, Apr 10, 2009 at 6:54 PM, Brian Whitman br...@echonest.com wrote:
 i have this version of solr running:

 Solr Implementation Version: 1.4-dev 747554M - bwhitman - 2009-02-24
 16:37:49

 and am trying to update a schema to support 8 code length metaphone instead
 of 4 via this (committed) issue:

 https://issues.apache.org/jira/browse/SOLR-813

 So I change the schema to this (knowing that I have to reindex)

        filter class=solr.PhoneticFilterFactory encoder=DoubleMetaphone
 inject=true maxCodeLength=8/

 But when I do queries fail with

 Error_initializing_DoubleMetaphoneclass_orgapachecommonscodeclanguageDoubleMetaphone__at_orgapachesolranalysisPhoneticFilterFactoryinitPhoneticFilterFactoryjava90__at_orgapachesolrschemaIndexSchema$6initIndexSchemajava821__at_orgapachesolrschemaIndexSchema$6initIndexSchemajava817__at_orgapachesolrutilpluginAbstractPluginLoaderloadAbstractPluginLoaderjava149__at_orgapachesolrschemaIndexSchemareadAnalyzerIndexSchemajava831__at_orgapachesolrschemaIndexSchemaaccess$100IndexSchemajava58__at_orgapachesolrschemaIndexSchema$1createIndexSchemajava425__at_orgapachesolrschemaIndexSchema$1createIndexSchemajava410__at_orgapachesolrutilpluginAbstractPluginLoaderloadAbstractPluginLoaderjava141__at_orgapachesolrschemaIndexSchemareadSchemaIndexSchemajava452__at_orgapachesolrschemaIndexSchemainitIndexSchemajava95__at_orgapachesolrcoreSolrCoreinitSolrCorejava501__at_orgapachesolrcoreCoreContainer$InitializerinitializeCoreContainerjava121