I have a problem with solr search: I have keyword, example: apache solr
reference, I index for apache, solr, reference.when I search with the
list keywords below:- apache solr reference - OK- apache - OK- solr - OK-
the same.But when the first keyword is not index, and other keywords are
index,
I think you have explained it perfectly on how the tag exclusion makes it a
different facet field and that no logic of default/invariants/appends would
be able to solve this
I went with the custom component approach.
Although a very hacky solution could be defining this in defaults:
str
Hi Michael,
I opened that ticket, and it looks like there is indeed a buffer or limit I
was exceeding. As per the ticket I guess the stream is cut off at that
limit, and is then malformed. I am using Tomcat, and since increasing some
limits on the connector, I haven't had any issues since. I'll
Prasi, as per the ticket I linked to earlier, I was running into GC
settings. May be worth investigating - and take a look at the GC settings
I'm running with in the ticket.
Cheers,
Chris
On 22 October 2013 10:25, Prasi S prasi1...@gmail.com wrote:
bq: ...three different files each with a
If I am not mistaken the most impressive improvement of Solr 4.0 compared
to previous versions was the Solr Cloud architecture.
What would be the next big thing in Solr 5.0 ?
Saar
We were trying to migrate to 4.5 from 4.0 and faced similar issue as well.
I saw the ticket raised by Chris and tried setting formdataUploadLimitInKB
to a higher value and which did not resolve this issue.
We use Solr 4.0.0 currently and no additional container settings are
required. But it is
Hello All,
Can some one explain me following snippet of SolrConfig.xml in terms of
Solr API (Java Psuedo Code) for better understanding.
like
*updateHandler class=solr.DirectUpdateHandler2*
* *
* UpdateLog*
* str dir=BLAH /*
*
Hi,
I'd really appreciate if you could give me some help understanding how
to tune the document cache.
My thoughts:
min values: max_results * max_concurrent_queries, as stated by
http://wiki.apache.org/solr/SolrCaching
how can I estimate max_concurrent_queries?
Hi,
I am migrating Solr 4.4 to Solr 4.5 and I have an issue in Solr.xml.
I my old Solr.xml I had some properties I am reusing for all my cores.
Furthemore I have some properties related to each individual core.
solr
property name=lucene.version value=LUCENE_40/
property name=store.fields
I think better understanding is a bit too vague. Is there a specific
problem you have? Your Jetty example would make sense if, for example, your
goal was to automatically generate solrconfig.xml from some other
configuration. But even then, you would probably use fillable templates and
don't need
Hi Sandeep,
You are quite likely below capacity with this current set-up:
http://wiki.apache.org/solr/SolrPerformanceProblems#OS_Disk_Cache
Few things for you to confirm:
1. Which version of Solr are you using?
2. The size of your index.
- Are fields stored? How much are these stored fields
Dear,
I would like to know if SOLR can do that:
I have a field named Assignee with values like:
Int Business Machines Corp
Int Business Mach Inc
I would like to have a result field in the schema.xml named
Norm_Assignee which contains
the translation with a lexical file:
Int Business Machines
I'm running 4.3 in solrcloud mode and trying to test index recovery, but
it's failing.
I have one shard, 2 replicas:
Leader: 10.159.8.105
Replica: 10.159.6.73
To test, I stopped the replica, deleted the 'data' directory and restarted
solr. Here is the replica's logging:
INFO - 2013-10-25
Maybe this can help you:
https://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters#solr.SynonymFilterFactory
-
Thanks,
Michael
--
View this message in context:
http://lucene.472066.n3.nabble.com/Normalized-data-during-indexing-tp4097750p4097752.html
Sent from the Solr - User mailing
I think what you are looking for is some kind of DTD/schema you can use to
see all the possible parameters in SolrConfig.xml, short answer, there
isn't one (currently) :(
jetty.xml has a DTD schema, and its XMLConfiguration format is inherently
designed to convert to code, so the list of possible
Sorry for the late reply, traveling yesterday...
That is odd on the surface of it. Try tailing out the solr log when you
fire your SolrJ
query to make sure the query even gets to Solr, perhaps something about how
you
try to route it goes to a place you don't intend...
I'm guessing it's just a
Well, there's no code here. But you can set up remote debugging fairly
easily from IntelliJ and Eclipse, so that's the firs thing I'd try.
Best,
Erick
On Wed, Oct 23, 2013 at 10:54 PM, Mingzhu Gao m...@adobe.com wrote:
Hi All ,
My custom analyser always hang when I click Analysis values
I don't know if this works for optimizing or not, but try attaching
distrib=false to the optimization request.
Hmmm, something that might be added to the UI, any admin UI guys listening?
:).
But I have to ask why you're optimizing anyway. Unless you have a very
specific reason,
it's usually
How many unique values are in the field? Solr has to create a counter
for each and every one of them, you may be blowing memory up. What
do the logs say?
Best,
Erick
On Thu, Oct 24, 2013 at 4:07 PM, Bruno Mannina bmann...@free.fr wrote:
Just a little precision: solr down after running my URL
Make it at least 4.5.1, which was just released!
Erick
On Thu, Oct 24, 2013 at 8:45 PM, Eric Grobler impalah...@googlemail.comwrote:
Hi Chris
Thank you for your response.
I will try to migrate to Solr 4.4 first!
Best regards
On Thu, Oct 24, 2013 at 10:44 PM, Chris Hostetter
Hi,
I'm using solr 4.3 and I have data in multiple cores which are different in
structure like (Core1 - col1 col2) (Core2 - col3 col4).
Now I would like to run a search query on both of the cores and in the end
to get a single result set from the 2 cores combines.
Please help me out is it
On 10/25/2013 3:31 AM, Saar Carmi wrote:
If I am not mistaken the most impressive improvement of Solr 4.0 compared
to previous versions was the Solr Cloud architecture.
What would be the next big thing in Solr 5.0 ?
There are a huge number of behind-the-scenes changes in 4.x compared to
3.x.
I hadn't thought about it before, but now I'm curious how
MMapDirectoryFactory plays into documentCache. Uwe,
are you listening? :) My _guess_ is that if you're using
MMapDirectoryFactory, the usefulness of the document
cache is lessened, kinda.
Since the documents are coming from essentially
They _should_ be just the same. The new solr.xml format just works
by moving the individual core tags and putting the properties
in individual core.properties, and removing the cores tag completely.
So the global properties you refer to should be unaffected. But I haven't
verified this personally.
This is usually an app-layer function, i.e. you fire two queries
from the app layer, one to each core and combine them
as you please.
One issue you will face is what combine them as you please
means. The scores re NOT comparable, so do you
interleave them? Group them in tabs? etc...
Best,
Erick
On 10/25/2013 7:48 AM, Erick Erickson wrote:
I hadn't thought about it before, but now I'm curious how
MMapDirectoryFactory plays into documentCache. Uwe,
are you listening? :) My _guess_ is that if you're using
MMapDirectoryFactory, the usefulness of the document
cache is lessened, kinda.
Thanks Chris and Sai. I was hoping to use the standard jetty configuration
(noting another thread on these forums indicating that it is the default and
supported container). but will migrate to tomcat of needed. Has anyone found a
workaround that works with the standard container?
We are
Hi,
We are trying to solve a business problem by performing solr more like this
query. We are able to perform the more like this search. We have a specific
use case that requires different boost on different match fields. Say i do
more like this based on fields title and description of products.
Hello,
Couple of questions related to using SPDY with solr.
1. Does anybody have experience running Solr on Jetty 9 with SPDY support -
and using Jetty Client (SPDY capable client) to talk to Solr over SPDY?
2. This is related to Solr - Cloud - inter node communication. This might
not be a
If it helps to clarify any, here's the full query:
/select
?
q=*:*
fq=type:ProjectGroup
fq={!join from=project_id_i to=project_id_im}user_id_i:65615 -role_id_i:18
type:UserRole
We have two Solr servers that were indexed from the same database. One of
the servers is running Solr 4.2, while the
Hi Erick,
I think it's a memory problem, I do my test on a little computer at home
(8Go Ram i3-2120 3.30Ghz 64bits)
and my database is very big 87M docs for 200Go size.
I thought SOLR could done statistic on only the query answer, so here on
around 3000 docs (around 6000 terms)
it's not so
Hi Michael,
thanks it sounds like I'm looking for
I need to investigate
Thanks a lot !
Le 25/10/2013 14:46, michael.boom a écrit :
Maybe this can help you:
https://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters#solr.SynonymFilterFactory
-
Thanks,
Michael
--
View this
In terms of adding or fixing documentation, the Installing Solr page
(https://cwiki.apache.org/confluence/display/solr/Installing+Solr)
includes a yellow box that says:
Solr ships with a working Jetty server, with optimized settings for
Solr, inside the example directory. It is recommended that
Hi Erik. thanks for your help.
I tried with solr.xml as follow
solr
property name=lucene.version value=LUCENE_40/
/solr
It fails with this exception
Caused by: org.apache.solr.common.SolrException: No system property or
default value specified for lucene.version value:${lucene.version}
Property substitution is controlled by the conf/solrcore.properties file.
The old solr.xml had a properties attribute to control where that file is
located, but that feature was removed from the new solr.xml (in Solr 4.4),
so the substitution properties must be in the solrcore.properties file
Synonyms would certainly let you index the field values consistently, but
you would not see the normalized values in query results, since results use
the stored, original value, not the final indexed term.
But you could use an update processor to map the terms that would be stored
and
Just to add to the “use jetty for Solr” argument - Solr 5.0 will no longer
consider itself a webapp and will consider the fact that Jetty is a used an
implementation detail.
We won’t necessarily make it impossible to use a different container, but the
project won’t condone it or support it and
Nothing of great significance that I have heard of, yet. Solr 4.x has lots
of ongoing significant improvements and will likely to have a lot more in
the months ahead. Ditto with Lucene improvements - lots of ongoing work to
enhance 4.x.
-- Jack Krupansky
-Original Message-
From:
I'm trying to put together a query in Solr that is becoming rather
complicated, and I'm not quite sure where to even start.
I'm building a directory, which for simplicity sake contains:
First Name
Last Name
Department Name (if faculty / staff)
User Types - faculty, staff, student - multivalued.
My e-book does describe all of the configuration elements of solrconfig in
some detail.
-- Jack Krupansky
-Original Message-
From: Amit Aggarwal
Sent: Friday, October 25, 2013 6:19 AM
To: solr-user@lucene.apache.org
Subject: Please explain SolConfig.xml in terms of SolrAPIs (Java
Right, but what if you have many properties being shared across multiple
cores.
That means you have to copy same properties in each individual
core.properties.
Is not this redundant data.
My main problem is I would like to keep several properties at solr level not
to core level.
Thanka a lot
Yeah , you caught it right Yes it was kid of Dtd .
Anyways thanks a lot for clearing my doubt ..
SOLVED .
On 25-Oct-2013 6:34 PM, Daniel Collins danwcoll...@gmail.com wrote:
I think what you are looking for is some kind of DTD/schema you can use to
see all the possible parameters in
I am currently using a separate core for indexing the autosuggest keywords.
Everything works fine except for one issue as below.
In index I have 2 entries
ring tone
ringtone
When users type in 'r' I display both ring tone and ringtone in auto suggest
list. I am trying to figure out a way to
Yes, it is unfortunate that the properties property was removed.
-- Jack Krupansky
-Original Message-
From: marotosg
Sent: Friday, October 25, 2013 12:52 PM
To: solr-user@lucene.apache.org
Subject: Re: Global User defined properties - solr.xml from Solr 4.4 to Solr
4.5
Right, but
Hi,
My search queries to solr are of the following nature,
(A OR B OR C) OR (X AND Y AND Z) OR ((ABC AND DEF) - XYZ)
What I am trying to achieve is when I fire the query the results returned
should be able to tagged with which part or the OR resulted in the result. In
case all three parts above
Can you file a JIRA issue?
- Mark
On Oct 25, 2013, at 12:52 PM, marotosg marot...@gmail.com wrote:
Right, but what if you have many properties being shared across multiple
cores.
That means you have to copy same properties in each individual
core.properties.
Is not this redundant data.
Saar,
The marriage with the big yellow elephant is a big deal. It changes the
scale.
Otis
Solr ElasticSearch Support
http://sematext.com/
On Oct 25, 2013 5:32 AM, Saar Carmi saarca...@gmail.com wrote:
If I am not mistaken the most impressive improvement of Solr 4.0 compared
to previous
But a lot of that big yellow elephant stuff is in 4.x anyway.
(Otis: I was afraid that you were going to say that the next big thing in
Solr is... Elasticsearch!)
-- Jack Krupansky
-Original Message-
From: Otis Gospodnetic
Sent: Friday, October 25, 2013 2:43 PM
To:
I'm rusty on SPDY. Can you summarize the benefits in Solr context? Thanks.
Otis
Solr ElasticSearch Support
http://sematext.com/
On Oct 25, 2013 10:46 AM, Vinay Pothnis poth...@gmail.com wrote:
Hello,
Couple of questions related to using SPDY with solr.
1. Does anybody have experience
: Search Department Name + Last Name every time
: include First Name if user type in (faculty, staff) or if another token
: matched last name.
I haven't tested this, but i think conceptually what you want is...
qq=richard frovarp
q={!maxscore v=$maxof}
maxof=({!dismax qf='lastname deptname'
Hi,
shouldn't there be a tag for the 4.5.1 release under
http://svn.apache.org/repos/asf/lucene/dev/tags/ ?
Or am I looking at the wrong place?
Regards,
André
I’ll look into it. I ran the command to create the tag, but perhaps it did not
‘take’ :)
- Mark
On Oct 25, 2013, at 3:56 PM, André Widhani andre.widh...@digicol.de wrote:
Hi,
shouldn't there be a tag for the 4.5.1 release under
http://svn.apache.org/repos/asf/lucene/dev/tags/ ?
Or am
I had created it in a ‘retired’ location. The tag should be in the correction
spot now.
Thanks!
- Mark
On Oct 25, 2013, at 4:04 PM, Mark Miller markrmil...@gmail.com wrote:
I’ll look into it. I ran the command to create the tag, but perhaps it did
not ‘take’ :)
- Mark
On Oct 25,
Thanks Erick! I will try specifying the distrib parameter.
As for why I am optimizing, well i do lots of delete by id and by query and
after a while about 30% of maxDocs are deletedDocs. On a 50G index that
means about 15G of space which I am trying to free by doing the
optimization.
it's
Thanks, Mark!
On 10/25/2013 2:29 PM, michael.boom wrote:
As for why I am optimizing, well i do lots of delete by id and by query and
after a while about 30% of maxDocs are deletedDocs. On a 50G index that
means about 15G of space which I am trying to free by doing the
optimization.
it's usually better NOT to
On Oct 24, 2013, at 6:37 AM, michael.boom my_sky...@yahoo.com wrote:
Any idea what is happening and why the core on which i wanted the
optimization to happen, got no optimization and instead another shard got
optimized, on both servers?
Sounds like a bug we should fix. If you don’t specify
: But when the first keyword is not index, and other keywords are index, solr
: can not query it.
: (example I search with: apacheee solr reference).
run your queries in a browser and add debug=query to see how your query
is being parsed.
you most likely are seeing the results of the parser
: I have an issue that is only coming on live environment. The DIH
: with URLDataSource is not working when the file size imported is large
: (i.e. 100kb above - which is not so large). If its large, it returns
: nothing (as seen in the Debug section of DataImport at Solr Admin).
are you sure
I am trying to figure out a way to standardize the stored values using a file
similar to synonyms.txt file.
For ex:
If I have 3 entries as below
name: apple banana
name: appleBanana
name: applebaNana
Mapping
apple banana, appleBanana, applebaNana= applebanana
I want to just have one entry
I (jokingly) propose we take it a step further and drop Java :)! I'm
getting tired of trying to scale GC'ing JVMs!
Tim
On 25/10/13 09:02 AM, Mark Miller wrote:
Just to add to the “use jetty for Solr” argument - Solr 5.0 will no longer
consider itself a webapp and will consider the fact that
Things have actually improved quite a bit in that area. There have been many
optimizations and additional ways to create large data structures off heap
added in recent releases.
Someday G1 might even help a bit.
- Mark
On Oct 25, 2013, at 7:20 PM, Tim Vaillancourt t...@elementspace.com wrote:
: I am trying to write some code to read rank data from external db, I saw
: some example done using database -
:
http://sujitpal.blogspot.com/2011/05/custom-sorting-in-solr-using-external.html,
...
: But is there any way to pass parameter or choose different database during
:
: I know that right now is not possible, but I think that it's a feature that
: can be implemented in a easy way creating a new status for shards.
:
: shardIsLastResort=true? On paper it seems like a simple addition, but I
: am not at familiar enough with the SolrCloud-code to guess if it is
Hi Chris,
My parser operand now is OR not AND, in schema.xml I define:
solrQueryParser defaultOperator=OR/
I use below config to index (integrate detecting language in other config).
analyzer type=index
tokenizer class=solr.ICUTokenizerFactory/
filter
: debug: {
: rawquerystring: display_add_level1: apache solr reference,
: querystring: display_add_level1: apache solr reference,
: parsedquery: display_add_level1:apache text:solr text:reference,
: parsedquery_toString: display_add_level1:apache text:solr
: text:reference,
note
Focus turned to the issue of Unable to open socket to
0:0:0:0:0:0:0:1/0:0:0:0:0:0:0:1:2181
That's apparently been problematic for others as well. It might be at root here.
I believe I am able to prove zookeeper is running by asking its
status, which reports at least something.
I moved the
Thanks for your enthusiasm reply. I will try it.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-search-in-case-the-first-keyword-are-not-index-tp4097699p4097862.html
Sent from the Solr - User mailing list archive at Nabble.com.
68 matches
Mail list logo