n-Gram, only works with queries of 2 letters

2008-06-23 Thread Norberto Meijome
hi there, my use case : I want to be able to match documents when only a partial word is provided. ie, searching for 'roc' or 'ock' should match documents containing 'rock'. As I understand, the way to solve this problem is to use the nGram tokenizer @ index time and the nGram analyser @

Re: n-Gram, only works with queries of 2 letters

2008-06-23 Thread Norberto Meijome
On Mon, 23 Jun 2008 16:23:55 +1000 Norberto Meijome [EMAIL PROTECTED] wrote: hi there, my use case : I want to be able to match documents when only a partial word is provided. ie, searching for 'roc' or 'ock' should match documents containing 'rock'. As I understand, the way to solve

Re: Can I specify the default operator at query time ?

2008-06-23 Thread Erik Hatcher
dismax is a different sort of parser, where AND/OR doesn't really make sense exactly. The mm (minimum match) parameter for dismax can be used to set things to be fully OR (mm=0) or fully AND (mm=100%) Erik On Jun 23, 2008, at 6:27 AM, Nikhil Chhaochharia wrote: I think this works

Re: n-Gram, only works with queries of 2 letters

2008-06-23 Thread Otis Gospodnetic
Hi, When you add debugQuery=true to the request, what does your query look like after parsing? Otis-- Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch - Original Message From: Norberto Meijome [EMAIL PROTECTED] To: solr-user@lucene.apache.org Sent: Monday, June 23,

Re: Wildcard search question

2008-06-23 Thread Erik Hatcher
Jon, You provided a lot of nice details, thanks for helping us help you :) The one missing piece is the definition of the text field type. In Solr's _example_ schema, bobby gets analyzed (stemmed) to bobbi[1]. When you query for bobby*, the query parser is not running an analyzer on

Re: Wildcard search question

2008-06-23 Thread Jon Drukman
Erik Hatcher wrote: Jon, You provided a lot of nice details, thanks for helping us help you :) The one missing piece is the definition of the text field type. In Solr's _example_ schema, bobby gets analyzed (stemmed) to bobbi[1]. When you query for bobby*, the query parser is not running

Re: Wildcard search question

2008-06-23 Thread Erik Hatcher
On Jun 23, 2008, at 4:45 PM, Jon Drukman wrote: Erik Hatcher wrote: Jon, You provided a lot of nice details, thanks for helping us help you :) The one missing piece is the definition of the text field type. In Solr's _example_ schema, bobby gets analyzed (stemmed) to bobbi[1]. When you

Re: Wildcard search question

2008-06-23 Thread Jon Drukman
Erik Hatcher wrote: No, because the original data is str name=nameBobby Gaza/str, so Bobby* would match, but not bobby*. string type (in the example schema, to be clear) does effectively no analysis, leaving the original string indexed as-is, case and all. [...] stemming and wildcard

Re: SpellCheckComponent: No file-based suggestions + Location issue

2008-06-23 Thread Grant Ingersoll
Inline below On Jun 21, 2008, at 5:28 PM, Ronald K. Braun wrote: 2. I'm having difficulty getting the FileBasedSpellChecker to work -- probably something silly on my part but given the above I thought there might be container differences that haven't been vetted. Here is my config:

Re: SpellCheckComponent: No file-based suggestions + Location issue

2008-06-23 Thread Grant Ingersoll
Hmm, if I add: lst name=spellchecker str name=classnamesolr.FileBasedSpellChecker/str str name=namefile/str str name=sourceLocationspellings.txt/str str name=characterEncodingUTF-8/str str name=indexDir./spellcheckerFile/str /lst To the examples declaration with

Re: Wildcard search question

2008-06-23 Thread Norberto Meijome
On Mon, 23 Jun 2008 14:23:14 -0700 Jon Drukman [EMAIL PROTECTED] wrote: ok well let's say that i can live without john/jon in the short term. what i really need today is a case insensitive wildcard search with literal matching (no fancy stemming. bobby is bobby, not bobbi.) what are my

Re: n-Gram, only works with queries of 2 letters

2008-06-23 Thread Norberto Meijome
On Mon, 23 Jun 2008 05:33:49 -0700 (PDT) Otis Gospodnetic [EMAIL PROTECTED] wrote: Hi, When you add debugQuery=true to the request, what does your query look like after parsing? BTW, I've tested same data + similar config using EdgeNGramTokenizer and this works properly - I can

Re: adding documents with json post

2008-06-23 Thread Cam Bazz
thanks a bunch. On Mon, Jun 23, 2008 at 4:39 AM, Otis Gospodnetic [EMAIL PROTECTED] wrote: Hi Cam, Yes, the various other formats are for responses only, as far as I'm aware. Otis -- Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch - Original Message From: Cam Bazz

Attempting dataimport using FileListEntityProcessor

2008-06-23 Thread mike segv
I'm trying to use the fileListEntityProcessor to add some xml documents to a solr index. I'm running a nightly version of solr-1.3 with SOLR-469 and SOLR-563. I've been able to successfuly run the slashdot httpDataSource example. My data-config.xml file loads without errors. When I attempt

Cost of having fieldTypes defined but not used

2008-06-23 Thread Norberto Meijome
Hi all, I'm curious , what is the cost (memory / processing time @ load? performance hit ? ) of having several unused fieldTypes defined in schema.xml ? cheers, B _ {Beto|Norberto|Numard} Meijome Egotism is the anesthetic that dulls the pain of stupidity. Frank Leahy

Re: Attempting dataimport using FileListEntityProcessor

2008-06-23 Thread Noble Paul നോബിള്‍ नोब्ळ्
hi , You have not registered any datasources . the second entity needs a datasource. Remove the dataSource=null and add a name for the second entity (good practice). No need for baseDir attribute for second entity . See the modified xml added below --Noble dataConfig dataSource

Re: Attempting dataimport using FileListEntityProcessor

2008-06-23 Thread Noble Paul നോബിള്‍ नोब्ळ्
Just extend XPathEntityProcessor override nextRow() after 100 return null. Use it as your processor --Noble On Tue, Jun 24, 2008 at 10:45 AM, Noble Paul നോബിള്‍ नोब्ळ् [EMAIL PROTECTED] wrote: Just extend XPathEntityProcessor override nextRow() after 100 . Use it as your processor return

Re: Attempting dataimport using FileListEntityProcessor

2008-06-23 Thread Shalin Shekhar Mangar
Hi Mike, Just curious to know the use-case here. Why do you want to limit updates to 100 instead of importing all documents? On Tue, Jun 24, 2008 at 10:23 AM, mike segv [EMAIL PROTECTED] wrote: That fixed it. If I'm inserting millions of documents, how do I control docs/update? E.g. if