hi there,
my use case : I want to be able to match documents when only a partial word is
provided. ie, searching for 'roc' or 'ock' should match documents containing
'rock'.
As I understand, the way to solve this problem is to use the nGram tokenizer @
index time and the nGram analyser @
On Mon, 23 Jun 2008 16:23:55 +1000
Norberto Meijome [EMAIL PROTECTED] wrote:
hi there,
my use case : I want to be able to match documents when only a partial word
is provided. ie, searching for 'roc' or 'ock' should match documents
containing 'rock'.
As I understand, the way to solve
dismax is a different sort of parser, where AND/OR doesn't really make
sense exactly. The mm (minimum match) parameter for dismax can be
used to set things to be fully OR (mm=0) or fully AND (mm=100%)
Erik
On Jun 23, 2008, at 6:27 AM, Nikhil Chhaochharia wrote:
I think this works
Hi,
When you add debugQuery=true to the request, what does your query look like
after parsing?
Otis--
Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
- Original Message
From: Norberto Meijome [EMAIL PROTECTED]
To: solr-user@lucene.apache.org
Sent: Monday, June 23,
Jon,
You provided a lot of nice details, thanks for helping us help you :)
The one missing piece is the definition of the text field type. In
Solr's _example_ schema, bobby gets analyzed (stemmed) to
bobbi[1]. When you query for bobby*, the query parser is not
running an analyzer on
Erik Hatcher wrote:
Jon,
You provided a lot of nice details, thanks for helping us help you :)
The one missing piece is the definition of the text field type. In
Solr's _example_ schema, bobby gets analyzed (stemmed) to
bobbi[1]. When you query for bobby*, the query parser is not running
On Jun 23, 2008, at 4:45 PM, Jon Drukman wrote:
Erik Hatcher wrote:
Jon,
You provided a lot of nice details, thanks for helping us help you :)
The one missing piece is the definition of the text field type.
In Solr's _example_ schema, bobby gets analyzed (stemmed) to
bobbi[1]. When you
Erik Hatcher wrote:
No, because the original data is str name=nameBobby Gaza/str, so
Bobby* would match, but not bobby*. string type (in the example
schema, to be clear) does effectively no analysis, leaving the original
string indexed as-is, case and all.
[...]
stemming and wildcard
Inline below
On Jun 21, 2008, at 5:28 PM, Ronald K. Braun wrote:
2. I'm having difficulty getting the FileBasedSpellChecker to work --
probably something silly on my part but given the above I thought
there might be container differences that haven't been vetted. Here
is my config:
Hmm, if I add:
lst name=spellchecker
str name=classnamesolr.FileBasedSpellChecker/str
str name=namefile/str
str name=sourceLocationspellings.txt/str
str name=characterEncodingUTF-8/str
str name=indexDir./spellcheckerFile/str
/lst
To the examples declaration with
On Mon, 23 Jun 2008 14:23:14 -0700
Jon Drukman [EMAIL PROTECTED] wrote:
ok well let's say that i can live without john/jon in the short term.
what i really need today is a case insensitive wildcard search with
literal matching (no fancy stemming. bobby is bobby, not bobbi.)
what are my
On Mon, 23 Jun 2008 05:33:49 -0700 (PDT)
Otis Gospodnetic [EMAIL PROTECTED] wrote:
Hi,
When you add debugQuery=true to the request, what does your query look like
after parsing?
BTW, I've tested same data + similar config using EdgeNGramTokenizer and this
works properly - I can
thanks a bunch.
On Mon, Jun 23, 2008 at 4:39 AM, Otis Gospodnetic
[EMAIL PROTECTED] wrote:
Hi Cam,
Yes, the various other formats are for responses only, as far as I'm aware.
Otis
--
Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
- Original Message
From: Cam Bazz
I'm trying to use the fileListEntityProcessor to add some xml documents to a
solr index. I'm running a nightly version of solr-1.3 with SOLR-469 and
SOLR-563. I've been able to successfuly run the slashdot httpDataSource
example. My data-config.xml file loads without errors. When I attempt
Hi all,
I'm curious , what is the cost (memory / processing time @ load? performance
hit ? ) of having several unused fieldTypes defined in schema.xml ?
cheers,
B
_
{Beto|Norberto|Numard} Meijome
Egotism is the anesthetic that dulls the pain of stupidity.
Frank Leahy
hi ,
You have not registered any datasources . the second entity needs a datasource.
Remove the dataSource=null and add a name for the second entity
(good practice). No need for baseDir attribute for second entity .
See the modified xml added below
--Noble
dataConfig
dataSource
Just extend XPathEntityProcessor override nextRow() after 100
return null. Use it as your processor
--Noble
On Tue, Jun 24, 2008 at 10:45 AM, Noble Paul നോബിള് नोब्ळ्
[EMAIL PROTECTED] wrote:
Just extend XPathEntityProcessor override nextRow() after 100 . Use
it as your processor
return
Hi Mike,
Just curious to know the use-case here. Why do you want to limit updates to
100 instead of importing all documents?
On Tue, Jun 24, 2008 at 10:23 AM, mike segv [EMAIL PROTECTED] wrote:
That fixed it.
If I'm inserting millions of documents, how do I control docs/update? E.g.
if
18 matches
Mail list logo