First excuse me if I do not use the correct terminology.
I have some records in a Solr document with a field called icDesc_en.
The contents of this field contains a sentences of two, e.g. 2.4l engine
automatic 5 speed, Left front door, electric windows, etc.
When I preform a search to
what's field type of icDesc_en?
See it in schema.xml in conf directory of your solr setup.
I guess it must be tokenized by tokenizer.
If that is the case than change the type of this field to string type.
By doing this tokens wouldn't be created and you will get desired results.
-Karan
On
Investigating, it looks that the payload.bytes property is where the problem
is.
payload.toString() outputs corrects values, but .bytes property seems to
behave a little weird:
public class CustomSimilarity extends DefaultSimilarity {
@Override
public float scorePayload(int doc, int
if you are interested here is the brief scratch of the possible hack:
- client adds special query ID parameter into request interruptID=DEADBEEF
- create InterruptionComponent and add into search handler list before
query component
- this component adds PostFilter which yields DelegatingCollector
Strange, is it really floats you are inserting as payload? We use payloads too
but write them via PayloadAttribute in custom token filters as float.
-Original message-
From:michael.boom my_sky...@yahoo.com
Sent: Tuesday 14th January 2014 11:59
To: solr-user@lucene.apache.org
Yes, it's float:
filter class=solr.DelimitedPayloadTokenFilterFactory encoder=float
delimiter=|/
The scenario is simple to replicate - default solr-4.6.0 example, with a
custom Similarity class (the one above) and a custom queryparser (again,
listed above).
I posted the docs in XML format (docs
Karan,
The field was a text type, which by experimentation I changed to string
and all was OK.
Thanks for your prompt reply.
David
--
View this message in context:
http://lucene.472066.n3.nabble.com/Facets-tp491p4111234.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi David,
As Karan suggested,your current icDesc_en is tokenized (understandably you
need to do that if you want to search on it in powerful way). So the
solution is create another field say icDesc_en_facet and define string as
the type (like Karan suggested) and then do this : copyField source=
I checked the index without any problems beeing found.
So it is not obvious, what is going wrong here while the index itself
looks okay.
Next step, updating java, is work in progress.
So I will come back after sorting out the java version as the cause for
the failing optimize.
The
thank you very much Mr. Sumit
--
View this message in context:
http://lucene.472066.n3.nabble.com/Can-I-define-the-copy-field-like-title-tp468p471.html
Sent from the Solr - User mailing list archive at Nabble.com.
I just wonder can I define copy field like this
copyField source=title_* dest=text/
instead of
copyField source=title_th dest=text/
copyField source=title_en dest=text/
millions thank you
Rachun
--
View this message in context:
Hi Markus,
Do you have any example/tutorials of your payloads in custom filter
implementation ?
I really want to get payloads working, in any way.
Thanks!
-
Thanks,
Michael
--
View this message in context:
I have a Java question, for a custom update processor I'm developing. It
takes an input field of the following format:
field:value;mvfield:value1;mvfield:value2
With an inner delimiter set to a colon and an outer delimiter set to a
semicolon, this results in two new fields going into the
Hi Shawn,
Solrj’s StrUtils.splitSmart() should do exactly what you want - in the first
pass, split on semicolon and don’t decode backslash escaping, and then in the
inner loop, use the same method to split on colons and decode backslash
escaping. I think :).
Steve
On Jan 14, 2014, at 10:07
Look at the StrUtils.splitSmart methods... the first variant treats
quotes specially,
the second variant doesn't (that's the one you probably want).
-Yonik
http://heliosearch.org -- off-heap filters for solr
On Tue, Jan 14, 2014 at 10:07 AM, Shawn Heisey s...@elyograg.org wrote:
I have a Java
I have a custom data import handler that creates an ExternalFileField from
a source that is different from the main index. If the import fails (in my
case, a connection refused in URLDataSource), I don't want to roll back any
uncommitted changes to the main index. However, this seems to be the
On 1/14/2014 8:20 AM, Steve Rowe wrote:
Solrj’s StrUtils.splitSmart() should do exactly what you want - in the first
pass, split on semicolon and don’t decode backslash escaping, and then in the
inner loop, use the same method to split on colons and decode backslash
escaping. I think :).
Hi all,
I'm almost new to SolR, and I have to make a improvment on a existing project,
but despite some hours of searching, I'm stuck.
We have an index containing products, which the user can search by reference,
or name.By now, when the user make a search by product name, the score is the
same
It's usually a mistake to try to tune at this level. The tf/idf
calculations _already_ take into account the field length (measured in
tokens) when scoring. Matches on shorter fields add more to the score
than matches on longer fields, which seems to be what you're looking
for.
That said, the
Are there any plans/tickets to allow for pluggable SolrConf and
CoreLocator? In my use case my solr.xml is totally static, i have a
separate dataDir and my core.properties are derived from a separate
configuration (living in ZK) but totally outside of the SolrCloud..
I'd like to be able to not
The problem is with the admin UI not following the XML include to find
entity so it found none. DIH itself does support XML include as I can
issue the DIH commands via HTTP on the included entities successfully.
Bill
On Mon, Jan 13, 2014 at 8:03 PM, Shawn Heisey s...@elyograg.org wrote:
On
Can someone shed some light on this?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Query-time-join-with-conditions-tp4108365p4111300.html
Sent from the Solr - User mailing list archive at Nabble.com.
The work done as part of new style solr.xml, particularly by
romsegeek should make this a lot easier. But no, there's no formal
support for such a thing.
There's also a desire to make ZK the one source of truth in Solr 5,
although that effort is in early stages.
Which is a long way of saying
I created SOLR-5630.
Although WildcardQuery is much much faster now with AutomatonQuery, it can
still result in slow queries when used in multiple keywords. From my
testing, I think I will need to disable all WildcardQuerys and only allow
PrefixQuery.
Peter
On Sat, Jan 11, 2014 at 4:17 AM,
Hi Steve,
I think this is a great idea. Currently the implementation of CoresLocator is
picked depending on the type of solr.xml you have (new- vs old-style), but it
should be easy enough to extend the new-style logic to optionally look up and
instantiate a plugin implementation.
Core
Hi Michael
Did you re-index after you register your custom similarity?
Ahmet
On Tuesday, January 14, 2014 4:36 PM, michael.boom my_sky...@yahoo.com wrote:
Hi Markus,
Do you have any example/tutorials of your payloads in custom filter
implementation ?
I really want to get payloads
Hi Ahmet,
Yes, I did, also tried various scenarios with the same outcome. I used the
stock example, with minimum customization ( custom similarity and query
parser ).
-
Thanks,
Michael
--
View this message in context:
Hi Sebastien,
I think you want to boost product names that start with query term, right? Or
in other words, boost if query term occurs within the first N words of a
document.
SpanFirstQuery seems an elegant way to do it.
https://issues.apache.org/jira/browse/SOLR-3925
Alternatively one can
Hi,
I'm planning to upgrade to Solr 4.6 to move from using Result Grouping to
CollapsingQParserPlugin. I'm currently using SolrCloud, couple of issues
with Result Grouping are :
1. Slow performance
2. Incorrect result count from ngroup
My understanding is that CollapsingQParserPlugin is aimed
It's last_index_time which is written after data import is finished
successfully. In case of an error, the file is unchanged and nothing is
updated.
-
Sumit Arora
--
View this message in context:
http://lucene.472066.n3.nabble.com/dataimport-properties-files-tp484p4111332.html
Sent
Shamik,
You still need to keep docs in the same group on the same shard with the
CollapsingQParserPlugin. If you use the group id as the shard-key with
SolrCloud's automatic document routing (
http://searchhub.org/2013/06/13/solr-cloud-document-routing/), the groups
will automatically end up on
Also, there are a number of bugs in the CollapsingQParserPlugin in Solr 4.6
that are resolved in Solr 4.6.1 which should be out soon.
Joel Bernstein
Search Engineer at Heliosearch
On Tue, Jan 14, 2014 at 10:00 PM, Joel Bernstein joels...@gmail.com wrote:
Shamik,
You still need to keep docs
Something is off but I'm not sure what. A couple of questions.
1) You mention updating the solr.xml. Did you the schema.xml?
2) Did you load only those 4 docs?
Joel
Joel Bernstein
Search Engineer at Heliosearch
On Mon, Jan 13, 2014 at 4:21 PM, Shamik Bandopadhyay sham...@gmail.comwrote:
Hi,
I just did a quick test with the 4 docs and got the proper result.
All I did was change the adskdedup field to adskdedup_s so it would load as
a dynamic string field. You can see the output below.
Can you provide more details on the exact steps you took?
{
responseHeader:{
status:0,
Just tried it with q=solr as well:
{
responseHeader:{
status:0,
QTime:1,
params:{
indent:true,
q:solr,
wt:json,
fq:{!collapse field=adskdedup_s}}},
response:{numFound:2,start:0,docs:[
{
id:SOLR1000,
name:Solr, the Enterprise Search
Hi Amit,
This excel sheet will help you estimating the index size.
size-estimator-lucene-solr.xls
http://lucene.472066.n3.nabble.com/file/n4111365/size-estimator-lucene-solr.xls
-
Sumit Arora
--
View this message in context:
you should be able to do the following
/ProfileCore/select?q=*:*fq={!join fromIndex=RssCore from=profile_id to=id
v=$rssQuery}rssQuery=(type:'RssEntry')
There is also a new join impl
https://issues.apache.org/jira/browse/SOLR-4787 which allows you to use fq
within join, which will support Nested
It becomes just more spooky.
The optimize-run this night was succesful.
Yesterday I did two things:
1. Checked the index without any result (no problems found).
2. I did an expungeDelete on the mentioned index.
So I have no idea what is going on here.
Btw: Java version is still the old 1.6.0.
Joel,
Thanks for the pointer. I went through your blog on Document routing, very
informative. I do need some clarifications on the implementation. I'll try
to run it based on my use case.
I'm indexing documents from multiple source system out of which a bunch
consist of duplicate content. I'm
Thanks Joel, I found the issue. It had to do with the schema definition for
adskdedup field. I had defined it as a text_general which was analyzing it
based on -. After I changed it to type string, it worked as expected.
Thanks for looking into this.
--
View this message in context:
40 matches
Mail list logo