you should try fq=Product:Electric Guitar
How do I handle facet values that contain whitespace? Say I have a field
Product that I want to facet on. A value for Product could be Electric
Guitar. How should I handle the white space in Electric Guitar during
indexing? What about when I
Hi All,
i am getting different results when i used with some escape keys..
for example:::
1) when i use this request
http://localhost:8080/solr/select?q=erlang!ericson
the result obtained is
result name=response numFound=1934 start=0
2) when the
Hi,
Is there a way with faceting or field collapsing to do the SQL equivalent of
SELECT count(distinct(field)) FROM index WHERE length(field) 0 AND
other_criteria
i.e. I'm only interested in the total count not the individual records
and counts.
Cheers,
Dan
We need more information about the the analyzers and tokenizers of the
default field of your search
Marco Martínez Bautista
http://www.paradigmatecnologico.com
Avenida de Europa, 26. Ática 5. 3ª Planta
28224 Pozuelo de Alarcón
Tel.: 91 352 59 42
2010/12/22 satya swaroop satya.yada...@gmail.com
try to copy the values (with copyfield) to a string field
Marco Martínez Bautista
http://www.paradigmatecnologico.com
Avenida de Europa, 26. Ática 5. 3ª Planta
28224 Pozuelo de Alarcón
Tel.: 91 352 59 42
2010/12/22 Peter Karich peat...@yahoo.de
you should try fq=Product:Electric Guitar
Hi all,
I am using solr in my web application for search purposes. However, i
am having a problem with the default behaviour of the solr search.
From my understanding, if i query for a keyword, let's say Laptop,
preference is given to result rows having more occurences of the search
Have you investigated 'field collapsing'? I believe that it is a least the
'DISTINCT' part.
Dennis Gearon
Signature Warning
It is always a good idea to learn from your own mistakes. It is usually a
better
idea to learn from others’ mistakes, so you do not have to make them
Hello.
I have a RAM problem during a optimize.
When is start an delta or full import, solr using only this ram which i
allocate to him.
eg.: java -jar -Xmx2g start.jar
when solr is fetching the rows from database the using of ram ist okay. But
when solr begin to otimize, solr want all of the
On Wed, Dec 22, 2010 at 3:09 PM, Hasnain hasn...@hotmail.com wrote:
[...]
From my understanding, if i query for a keyword, let's say Laptop,
preference is given to result rows having more occurences of the search
keyword Laptop in the field name. This, however, is producing
undesirable
maybe i set my chache in solrconfig.xml to high ? why now i see das the cache
very high in server.
--
View this message in context:
http://lucene.472066.n3.nabble.com/hole-RAM-using-by-solr-during-Optimize-tp2130482p2130490.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi,
First of all thanks for replying.
Secondly, maybe i wasn't clear enough in my original post regarding what was
required and what has been implemented.
In my schema, i have another field by the name of Category and, for
example's sake, let's assume that my application supports only two
facets=truefacet.field=field // SELECT count(distinct(field))
fq=field:[* TO *] // WHERE length(field) 0
q=other_criteriaAfq=other_criteriaB// AND other_criteria
advantage: you can look into several fields at one time when adding
another facet.field
disadvantage: you get the counts splitted by
Hi,
Is there any support for glob in the 'fl' param. This would be very useful
in case of retrieving dynamic fields. I have read the wiki for
FieldAliasesAndGlobsInParams. Is there any related patch?
Thanks for any pointers,
Samarth
Hello Erick,
Could you kindly give a hand on my problem. Any ideas, hints,
suggestions are highly appreciated. Many thanks
1. The problem: Solr index directory '/solr/data/index' doesn't exist.
Creating new index...
2. Some other info.:
- use the solr example 1.4.1
- Geronimo 2.1.6
- solr
I want to run delta-import in Crontab but don't know how.
I used php file in Crontab before, like:
command: php /home/user/public_html/auto.php
I tried:
command:
/home/user/public_html/solr/apache-solr-1.4.1/example/example-DIH/solr/db/dataimport?command=delta-import
It didn't work.
The url
Hi,
you can use wget if available on your server, e.g. command
wget --quiet
'http://181.163.64.228:8983/solr/db/dataimport?command=delta-import'
Cheers,
Stefan
Am 22.12.2010 12:31, schrieb Ruixiang Zhang:
I want to run delta-import in Crontab but don't know how.
I used php file in Crontab
Thanks for your quick reply. I couldn't find the wget on my server. Do you
know where it should be located or how I can check if I have it on my
server? If not, can I install one?
Thanks
On Wed, Dec 22, 2010 at 3:38 AM, Stefan Moises moi...@shoptimax.de wrote:
Hi,
you can use wget if
Just call wget http://www.somedomain.com on the console to see if it is
available...
Depends on your distro where it is installed and how to install it... I
have mine in /usr/bin/wget
Alternatively, use lynx or curl as command, e.g.
curl --silent
1) Thanks for this update. I have to use 'WhiteSpaceTokenizer'
2) I have to suggest the whole query itself (Say name or title)
3) Could you please let me know if there is a way to find the evicted docs?
4) Yes, we are seeing improvement in the response time if we optimize. But
still for some
Hi,
I got small problem with DIH for SolrCloud. I have specified my
dataSource settings in the seperated file: data-config.xml in the conf
folder (same folder where schema.xml and solrconfig are placed).
When I try importing my data from DB table for indexing I receive the
following problem:
Tried to checkout lucene/solr and setup projects and classpath in eclipse -
there seems to be circular dependency between modules - this is not
possible/allowed in maven built project, would require refactoring.
Regards,
Stevo.
On Wed, Dec 8, 2010 at 1:42 PM, Stevo Slavić ssla...@gmail.com
This won't actually give you the number of distinct facet values, but will give
you the number of documents matching your conditions. It's more equivalent to
SQL without the distinct.
There is no way in Solr 1.4 to get the number of distinct facet values.
I am not sure about the new features
Well, SolrCloud is not yet fully specified for the indexing side - more work
remains.
But my point is that the architecture for should be ZK based.
I added a new jira issue to flesh out a strategy for SolrCloud controlled
distributed indexing in SOLR-2293
Perhaps you should open a JIRA issue
The phrase solution works as does escaping the space with a backslash:
fq=Product:Electric\ Guitar ... actually a lot of characters need to be escaped
like this (amperstands and parenthesis come to mind)...
I assume you already have this indexed as string, not text...
James Dyer
E-Commerce
Hi,
In a test set up i have a master and slave in the same JVM but different cores.
Of course i'd like to replicate configuration files and include some via
XInclude.
The problem is the href path; it's can't use properties and is relative to the
servlet container.
Here's the problem, i also
On Wed, Dec 22, 2010 at 9:53 AM, Dyer, James james.d...@ingrambook.com wrote:
The phrase solution works as does escaping the space with a backslash:
fq=Product:Electric\ Guitar ... actually a lot of characters need to be
escaped like this (amperstands and parenthesis come to mind)...
One
What you want to ask? When this problem arises.? Is it when you try to index
to solr? What are the commands that you are running? Which version of solr(
1.4.1?).
On Wed, Dec 22, 2010 at 5:49 PM, Bac Hoang [via Lucene]
I'm using SOLR 1.4.1 with SOLR-1553 applied (edismax query parser). I'm
experiencing inconsistent behavior with terms grouped in parenthesis.
Sometimes they are AND'ed and sometimes OR'ed together.
1. q=Title:(life)defType=edismax 285 results
2. q=Title:(hope)defType=edismax 34 results
On Wed, Dec 22, 2010 at 4:55 PM, Bac Hoang bac.ho...@axonactive.vn wrote:
Hello Erick,
Could you kindly give a hand on my problem. Any ideas, hints, suggestions
are highly appreciated. Many thanks
1. The problem: Solr index directory '/solr/data/index' doesn't exist.
Creating new index...
Hello,
My main (full text) index contains the terms www, sometest, com, which
is intended and correct.
My spellcheck index contains the term www.sometest.com. which is also
intended and correct.
However, when querying the spellchecker using the query www.sometest.com,
I get the suggestion
Hi,
I have a solar index that has thousands of records, the title is one of the
solar fields, and I would like to query for title values that are less than
50 characters long. Is there a way to construct the Solr query to provide
results based on the character length?
thank you very much!
Check the analyzer of the field you defined for queryAnalyzerFieldType which is
configured in the search component.
On Wednesday 22 December 2010 16:32:18 Sebastian M wrote:
Hello,
My main (full text) index contains the terms www, sometest, com,
which is intended and correct.
My
Hello Anurag,
The specific problem I faced when started solr in Geronimo
(http://{server}:{port}/solr) is /solr/data/index could not be found, then solr
tried to create that folder but failed, even permission is granted.
More detail got from the log: Solr index directory '/solr/data/index'
In my experience, that should work fine. Facetting in 1.4 works fine on
multi-valued fields, and a duplicate value in the multi-valued field
shouldn't be a problem.
On 12/22/2010 2:31 AM, Andy wrote:
If I put duplicate values into a multiValued field, would that cause any issues?
For example
On Wed, Dec 22, 2010 at 9:06 PM, Giri giriprak...@gmail.com wrote:
Hi,
I have a solar index that has thousands of records, the title is one of the
solar fields, and I would like to query for title values that are less than
50 characters long. Is there a way to construct the Solr query to
Hi and thanks for your reply,
My searchComponent is as such:
searchComponent name=spellcheck class=solr.SpellCheckComponent
str name=queryAnalyzerFieldTypetextSpell/str
...
/searchComponent
And then in my schema.xml, I have:
fieldType name=textSpell class=solr.TextField
Another technique, which works great for facet fq's and avoids the need
to worry about escaping, is using the field query parser instead:
fq={!field f=Product}Electric Guitar
Using the field query parser avoids the need for ANY escaping of your
value at all, which is convenient in the
Huh, does !term in 4.0 mean the same thing as !field in 1.4? What you
describe as !term in 4.0 dev is what I understand as !field in 1.4 doing.
On 12/22/2010 10:01 AM, Yonik Seeley wrote:
On Wed, Dec 22, 2010 at 9:53 AM, Dyer, Jamesjames.d...@ingrambook.com wrote:
The phrase solution works
On Dec 22, 2010, at 09:21 , Jonathan Rochkind wrote:
This won't actually give you the number of distinct facet values, but will
give you the number of documents matching your conditions. It's more
equivalent to SQL without the distinct.
There is no way in Solr 1.4 to get the number of
I have two request handlers set up something like this:
requestHandler name=Keyword_SI class=solr.SearchHandler
lst name=defaults
str name=defTypeedismax/str
float name=tie0.01/float
str name=qfTitle^130 Features^110 Edition^100 CTBR_SEARCH^90
Hi,
I am looking into using a multi core configuration to allow us to
fully rebuild our index while still applying updates.
I have two cores main-core and rebuild-core. I push the whole dataset
into the rebuild core, during which time I can happily keep pushing
updates into the main-core. Once
Stevo,
You may be interested in LUCENE-2657
https://issues.apache.org/jira/browse/LUCENE-2657, which provides full POMs
for Lucene/Solr trunk.
I don't use Eclipse, but I think it can use POMs to bootstrap project
configuration. (I know IntelliJ can do this.)
Steve
-Original
No good way. At indexing time, I'd just store the number of chars in the
title in a field of it's own. You can possibly do that solely in
schema.xml with clever use of analyzers and copyField.
Solr isn't an rdbms. Best to de-normalize at index time so what you're
going to want to query is
Well, that's true -- you can get the total number of facet values if you
ALSO are willing to get back every facet value in the response.
If you've got a hundred thousand or so unique facet values, and what you
really want is just the _count_ without ALSO getting back a very large
response
Hi guys,
There's one more thing to get this code to work as I need I just found
out...
Im now using: q=title_search:hort*defType=lucene
as iorixxx suggested.
it works good BUT, this query doesnt find results if the title in DB is
Hortus supremus
I tried adding some tokenizers and filters to
: Huh, does !term in 4.0 mean the same thing as !field in 1.4? What you
: describe as !term in 4.0 dev is what I understand as !field in 1.4 doing.
There is a subtle distinction between {!field}, {!raw}, and {!term} which
i attempted to explain on slides 26 and 43 in this presentation...
: If I put duplicate values into a multiValued field, would that cause any
issues?
:
: For example I have a multiValued field Color. Some of my documents
: have duplicate values for that field, such as: Green, Red, Blue, Green,
: Green.
:
: Would the above (having 3 duplicate Green) be the
Did you reindex after you changed your analyzers?
On 12/22/2010 12:57 PM, PeterKerk wrote:
Hi guys,
There's one more thing to get this code to work as I need I just found
out...
Im now using:q=title_search:hort*defType=lucene
as iorixxx suggested.
it works good BUT, this query doesnt find
No, one cannot ignore the schema. If you try to add a field not in the
schema you get
an error. One could, however, use any arbitrary subset
of the fields defined in the schema for any particular #document# in the
index. Say
your schema had fields f1, f2, f3...f10. You could have fields f1-f5 in
PositionIncrementGap for multiValued fields is, perhaps, the most
interesting
difference. One of the drivers here is, say, indexing across some boundary
that you don't want phrases or near clauses to match. For instance, say you
have text with
sentences, and your requirement is that phrases don't
I'm open to cores, if it's the faster(indexing/querying/keeping mentally
straight) way to do things.
But from what you say below, the eventual goal of the site would mean either
100
extra 'generic' fields, or 1,000-100,000's of cores.
Probably cores is easier to administer for security and
Certainly did!
Why, are you saying this code is correct as-is?
--
View this message in context:
http://lucene.472066.n3.nabble.com/full-text-search-in-multiple-fields-tp1888328p2133022.html
Sent from the Solr - User mailing list archive at Nabble.com.
: Hoss, I think the use case being asked about is specifically doing a
: facet.sort though, for cases where you actually do want to sort facet values
: with facet.sort, not sort records -- while still presenting the facet values
: with original case, but sorting them case insensitively.
Ah yes
So problem may be that index folder was not able to create. So try to
check the conf folder where solconfig.xml schema.xml resides. Also u may
try to index using $java -jar post.jar *.xml files.
You may try different version like 1.3.0 or 1.4.0 to test what is wrong. It
sometimes happens that
I've just done a bit of playing here, because I've spent a lot of time
reading the SolrReplication wiki page[1], and have often wondered how
some features interact.
Unfortunately, if you specify str name=enablefalse/str in your
replication request handler for your master, you cannot re-enable it
A dynamic field just means that the schema allows any field with a
name matching the wildcard. That's all.
There is no support for referring to all of the existing fields in the
wildcard. That is, there is no support for *_en:word as a field
search. Nor is there any kind of grouping for facets.
On 12/22/2010 8:25 AM, Dyer, James wrote:
I'm using SOLR 1.4.1 with SOLR-1553 applied (edismax query parser). I'm
experiencing inconsistent behavior with terms grouped in parenthesis.
Sometimes they are AND'ed and sometimes OR'ed together.
1. q=Title:(life)defType=edismax 285 results
2.
On 12/22/2010 2:56 AM, stockii wrote:
Hello.
I have a RAM problem during a optimize.
When is start an delta or full import, solr using only this ram which i
allocate to him.
eg.: java -jar -Xmx2g start.jar
when solr is fetching the rows from database the using of ram ist okay. But
when solr
Certainly did!
Why, are you saying this code is correct as-is?
Yes, the query q=title_search:hort*defType=lucene should return documents
having Hortus supremus in their title field with the configurations you send
us.
It should exists somewhere in the result set, if not in the top 10.
Try
:
: 1) My categories are stored in database as coded numbers instead of
: fully spelled out names. For example I would have a category of 2/7
: and a lookup dictionary to convert 2/7 into NonFic/Science. How do I
: do such lookup in DIH?
My advice: don't.
I thought i mentioned this in that
Ok, I was trying to hide the actual name of the location, because I dont want
it to get indexed by search engines AND its a bit of a weird name :p
The name of the location in the database is: Museumrestaurant De Pappegay
Anyway, here it is, I executed the queries you gave me, and this is the
--- On Wed, 12/22/10, satya swaroop satya.yada...@gmail.com wrote:
From: satya swaroop satya.yada...@gmail.com
Subject: Different Results..
To: solr-user@lucene.apache.org
Date: Wednesday, December 22, 2010, 10:44 AM
Hi All,
i am getting
different results when i used with some
The name of the location in the database is:
Museumrestaurant De Pappegay
What was the wildcard query for this?
I want to sort results as follows
- highest membervalue (float) on top.
- within those results I want to sort the items that share the same position
on the user rating (integer), once again, highest rating on top
- and within those results I want to sort the items that share the same
position on
Shawn,
Thank you for the reply. The URL you gave was helpful and Smiley Pugh even
more so. On Smiley Pugh page 140, they indicate that mm=100% using dismax is
analogous to Standard's q.op=AND. This is exactly what I need.
However...testing with these queries and edismax, I get different #
I have products, each has a specific Product ID.
For certain products such as Shirts, there are also extra fields such as
Size and Color.
Right now I define both Size and Color as multiValued fields. And when I
have a Shirt of Size M and Color white, I just put M in Size and white in
Color.
--- On Thu, 12/23/10, PeterKerk vettepa...@hotmail.com wrote:
From: PeterKerk vettepa...@hotmail.com
Subject: Sorting results on MULTIPLE fields, not showing expected order
To: solr-user@lucene.apache.org
Date: Thursday, December 23, 2010, 1:01 AM
I want to sort results as follows
-
Mmmm, this is strange:
When I do:
q=title_search:Pappegay*defType=luceneq=*:*fl=id,title
nothing is found.
but if I do:
q=title_search:PappegaydefType=luceneq=*:*fl=id,title
the location IS found.
I do need a wildcard though, since users may also search on parts of the
title (as described
Wow, you're fast :)
But that indeed did the trick, thanks!
--
View this message in context:
http://lucene.472066.n3.nabble.com/Sorting-results-on-MULTIPLE-fields-not-showing-expected-order-tp2133959p2134000.html
Sent from the Solr - User mailing list archive at Nabble.com.
--- On Wed, 12/22/10, Chris Hostetter hossman_luc...@fucit.org wrote:
: 2) Once I have the fully spelled out category path such
as
: NonFic/Science, how do I turn that into 0/NonFic
: 1/NonFic/Science using the DIH?
I don't have any specific suggestions for you -- i've never
tried it
When I do:
q=title_search:Pappegay*defType=luceneq=*:*fl=id,title
nothing is found.
but if I do:
q=title_search:PappegaydefType=luceneq=*:*fl=id,title
the location IS found.
I do need a wildcard though, since users may also search on
parts of the
title (as described earlier in
When I do:
q=title_search:Pappegay*defType=luceneq=*:*fl=id,title
nothing is found.
This is expected since you have lowercase filter in your index analyzer.
Wildcard searches are not analyzed. So you need to lowercase your query on
client side.
Oeps, sloppy, was a copy paste error.
I now have:
WORKING:
http://localhost:8983/solr/db/select/?indent=onq=title_search:PappegaydefType=lucenefl=id,title
NOT WORKING:
http://localhost:8983/solr/db/select/?indent=onq=title_search:Pappegay*defType=lucenefl=id,title
--
View this message in
I think I'm just going to have to have my partner and I play with both cores
and
dynamic fields.
If multiple cores are queried, and the schemas match up in order and postion
for
the base fields, the 'extra' fields in the different cores just show up in the
result set with their field names?
Hmmm. find evicted docs? If you mean find out how many docs are deleted,
look
on the admin schema browser page and the difference between MaxDoc and
NumDocs
is the number of deleted documents.
You say for some queries the QTime is more than 8 secs. What happens if
you
re-run that query a bit
I want to print the highlighting descriptions:
{responseHeader:{status:0,
QTime:2,params:{hl.fl:description,json.wrf:jsonp1293069622009,wt:json,q:target,hl:true}},response:{numFound:7945,start:0,maxScore:6.9186745,docs:[{description:target,url:target,id:269653,score:6.9186745},{description:Target
(10/12/23 11:56), Ruixiang Zhang wrote:
I want to print the highlighting descriptions:
{responseHeader:{status:0,
Thanks Koji. Actually my question is:
We can use response.response.docs[i].description
to print the description in docs.
What expression should we use to print the description in highlighting?
(10/12/23 14:10), Ruixiang Zhang wrote:
Thanks Koji. Actually my question is:
We can use response.response.docs[i].description
to print the description in docs.
What expression should we use to print the description in highlighting?
Ruixiang,
I cannot understand your question. Is it Solr
Hi Koji
I figured it out. I can use
response.highlighting[response.response.docs[0].id].description[0]
to print the description in highlighting.
(Actually, it's not a solr question, sorry for that.)
Thanks
Ruixiang
On Wed, Dec 22, 2010 at 10:05 PM, Koji Sekiguchi k...@r.email.ne.jp wrote:
80 matches
Mail list logo