Hi,
Any suggestion will be really helpful. Kindly provide your inputs.
Thanks,
Modassar
On Thu, Apr 16, 2015 at 4:27 PM, Modassar Ather modather1...@gmail.com
wrote:
Hi,
I have a setup of 5 node SolrCloud (Lucene/Solr version 5.1.0) without
replicas. When I am executing complex and large
Optimize will be distributed to all shards/replicas.
I believe reload will only reload the specific core. For reloading the
complete collection use the Collections API:
https://cwiki.apache.org/confluence/display/solr/Collections+API
On Thu, Apr 16, 2015 at 5:15 PM, Vincenzo D'Amore
II. Is there a way to use the stats.calcdistinct functionality and only
return the countDistinct portion of the response and not the full list of
distinct values -- as provided in the distinctValues portion of the
response. In a field with high cardinality the response size becomes too
Hi,
we have migrated Solr from 5.0 do 5.1 and we can't search now, we have a
ERROR for SolrCore like in subject. I can't get any info through Google.
Please, can someone help what is going on?
Thanks,
Pavel
--
View this message in context:
I don't think there is any Collection level support at this point in the
Solr admin UI. Whatever you do via the UI would be core level, unless I'm
forgetting something.
On Thu, Apr 16, 2015 at 5:15 PM, Vincenzo D'Amore v.dam...@gmail.com
wrote:
Hi all,
I have a solrcloud cluster with 3 server
Hi All,
Good Morning!!
For SolrCloud deployment, for indexing data through SolrJ, which is the
preferred / correct SolrServer class to use? HttpSolrServer of
CloudSolrServer? In case both can be used, when to use which? Any help
please.
Thanks Regards
Vijay
--
The contents of
Hi
I have enabled spellcheck but not getting any suggestions
withincorrectly spelled keywords.
I added the spellcheck into the/select request handler.
What steps did I miss out?
spellcheck list in return result:
lst name=spellcheck
lst name=suggestions/
/lst
solrconfig.xml:
requestHandler
Thank you for the reply.
Out schema is:
1) Index real-time (on separate machine).
2) NRT index becomes large.
3) Copy NRT index on other machine.
3) Merge NRT-made indexes with large (all-the-time) index.
4) Remove NRT index (until now it was available for searching).
At the end we have big,
Hi Norgorn,
I think there is no ready-made tool out of the box, but you have the spare
parts in the MapreduceIndexerTool :-)
With little effort you can decouple the index merging component from
MRIndexerTool and use based on the needs.
I did the same.
On Fri, Apr 17, 2015 at 10:40 AM, Norgorn
If you're using SolrCloud then you should use CloudSolrServer as it is
able to abstract / hide the interaction with the cluster. HttpSolrServer
communicates directly with a Solr instance.
Best,
Andrea
On 04/17/2015 10:59 AM, Vijay Bhoomireddy wrote:
Hi All,
Good Morning!!
For
Hello,
What is the best way to search in a field ignoring accents?
The field has the type:
fieldType name=text_general_edge_ngram
class=solr.TextField positionIncrementGap=100
analyzer type=index
Hi All,
I have solr synonyms stored in multiple files as defined in the schema:
!ENTITY sinonimi_freeling
Vincenzo D'Amore v.dam...@gmail.com wrote:
I have a SolrCloud cluster with 3 server, I would like to use stats.facet,
but this feature is available only if I upgrade to 4.10.
May I simply redeploy new solr cloud version in tomcat or should reload all
the documents?
There are other drawbacks?
Shouldn't you specify a spellcheck.dictionary in your request handler?
Best regards,
Elisabeth
2015-04-17 11:24 GMT+02:00 Derek Poh d...@globalsources.com:
Hi
I have enabled spellcheck but not getting any suggestions withincorrectly
spelled keywords.
I added the spellcheck into the/select
On 4/17/2015 6:02 AM, Vincenzo D'Amore wrote:
I have solr synonyms stored in multiple files as defined in the schema:
!ENTITY sinonimi_freeling
Hi Pedro,
solr.ASCIIFoldingFilterFactory is one way to remove diacritics.
Confusion comes from EdgeNGram, why do you need it?
Ahmet
On Friday, April 17, 2015 1:38 PM, Pedro Figueiredo
pjlfigueir...@criticalsoftware.com wrote:
Hello,
What is the best way to search in a field ignoring
Hi Ahmet,
Yes... the EdgeNGram is what produces those results...
I need it to improve the search by name by the applications users.
Thanks.
Pedro Figueiredo
Senior Engineer
pjlfigueir...@criticalsoftware.com
M. 934058150
Rua Engº Frederico Ulrich, nº 2650 4470-605 Moreira da Maia, Portugal
On 4/16/2015 2:07 PM, Steven White wrote:
In my case, I have to deploy Solr on Windows, AIX, and Linux (all server
edition). We are a WebSphere shop, moving away from it means I have to
deal with politics and culture.
You *can* run Solr 5.0 (and 5.1) in another container, just like you
could
Thanks for the fast turnaround, you beat me to opening the Jira and fixed it
too! Much appreciated.
Thanks,
Will
From: Tomás Fernández Löbbe tomasflo...@gmail.com
Sent: Thursday, April 16, 2015 10:26 PM
To: solr-user@lucene.apache.org
Subject: Re: Range
On 4/16/2015 2:27 PM, Oded Sofer wrote:
The issue is the firewall setting needed for the cloud. We do not want to
open all nodes to all others nodes. However, we found that add-index to a
specific node tries to access all other nodes though we set it to index
locally on that node only.
Hi,
sounds like you hit a Full GC. Check your GC.log.
Ugo
On 17 Apr 2015 08:24, Modassar Ather modather1...@gmail.com wrote:
Hi,
Any suggestion will be really helpful. Kindly provide your inputs.
Thanks,
Modassar
On Thu, Apr 16, 2015 at 4:27 PM, Modassar Ather modather1...@gmail.com
Hi,
this morning I have optimised my SolrCloud cluster (3 instances).
I have many collections, all are in shard and replica for each node.
At the end of optimisation task (about 10 minutes) all cores are optimised
on every node.
How can be sure than also reload affects all the cores?
On Fri,
On 4/17/2015 7:21 AM, Vincenzo D'Amore wrote:
this morning I have optimised my SolrCloud cluster (3 instances).
I have many collections, all are in shard and replica for each node.
At the end of optimisation task (about 10 minutes) all cores are optimised
on every node.
How can be sure than
And for this example what filter should I use?
Filter by edr should give the result Pedro
The NGram create tokens starting at the beginning or the ending, and in the
middle?
Thanks!
Pedro Figueiredo
Senior Engineer
pjlfigueir...@criticalsoftware.com
M. 934058150
Rua Engº Frederico Ulrich,
Hi Jainam,
One workaround is to use facet.query and frange query parser.
facet.query={!frange l=50 u=100}field(price)
Ahmet
On Thursday, April 16, 2015 1:01 PM, jainam vora jainam.v...@gmail.com wrote:
Hi,
I am using external field for price field since it changes frequently.
generate
Solr/Lucene are supposed to _always_ read one major version back. Thus
your 4.10 should be able to read indexes produced all the way back to
(and including) 3.x. Sometimes experimental formats are excepted.
In your case you should be fine since you're upgrading from 4.8..
As always, though, I'd
I like the first way. It matches how elasticsearch does it
http://www.elastic.co/guide/en/elasticsearch/reference/1.x/search-aggregations-bucket-range-aggregation.html
Can we specify explicit ranges in Solr now like we can in elasticsearch?
I do like how Solr's version of aggs can be much
I prefer the second way. I find it more readable and shorter.
Thanks for making Solr even better ;)
From: Yonik Seeley ysee...@gmail.com
Sent: Friday, April 17, 2015 12:20 PM
To: solr-user@lucene.apache.org
Subject: Re: JSON Facet Analytics API in Solr
Personally I find the second form easier to read. The second level of
nesting in the first example confuses me at first glance.
I don't have a really strong preference here, but I vote for the second form.
On Fri, Apr 17, 2015 at 9:20 AM, Yonik Seeley ysee...@gmail.com wrote:
Does anyone have
Hi,
df and q.op are the ones you are looking for.
You can define them in defaults section.
Ahmet
On Friday, April 17, 2015 9:18 PM, Bruno Mannina bmann...@free.fr wrote:
Dear Solr users,
Since today I used SOLR 5.0 (I used solr 3.6) so i try to adapt my old
schema for solr 5.0.
I have two
Thanks for getting back. Something like that crossed my mind but I checked
the values on the way into SolrJ SolrInputDocument match the values printed in
the Admin Query interface and they both match the expected value in the error
message exactly.
Besides the difference is only in the last
Here's another data point. To work around this issue, I am converting all
non-null _version_ values to the constant 1 on the way into Solr. As a result,
updates work fine.
Immediately after the update+commit, a /select?q=*:* returns the _version_
value of 1498715798795976700 for id ==
Pedro:
For your example, don't use EdgeNgrams, use just NGrams. That'll index
tokens like
(in the 2gram case) pe er dr ro and searching against edr would look
for ed dr. which would match.
However, this isn't in line with your first example where you got
results you didn't expect. You'll have to
Thanks for your answers, I looked at changes and we don't use
DocValuesFormat.
The question is, if I upgrade the SolrCloud version to 4.10, should I
reload entirely all documents?
Is there a binary compatibility between these two versions reading the
solar home?
On Fri, Apr 17, 2015 at 7:04 PM,
Dear Solr users,
Since today I used SOLR 5.0 (I used solr 3.6) so i try to adapt my old
schema for solr 5.0.
I have two questions:
- how can I set the defaultSearchField ?
I don't want to use in the query the df tag because I have a lot of
modification to do for that on my web project.
- how
you still haven't provided any details on what your client code looks like
-- ie: what code is talking to solr? what response format is it asking
for? is it JSON? what is parsing that JSON?
as for the admin UI: if you are looking at a JSON response in the Query
screen of the Admin UI, then
The core admin MERGEINDEXES will work for you I'm pretty sure. You
copy the NRT index over to the all-the-time box. MERGEINDEXES just
takes the path to the index you want to add to the existing core.
Note the warnings in the reference guide about taking care that the
indexes aren't changing and
Hi Pedro,
Requirement of Filter by edr should give the result Pedro can be done
expanding terms at index time only.
You can remove the ngram filter from query analyzer.
But remember that ngram filter produces a lot of tokens. Try it on analysis
page.
Regarding starting at the beginning or the
Additionally when indexing, CloudSolrServer collects up the documents
for each shard and routes them to the leader for that shard, moving
that processing away from whatever node you happen so contact using
HttpSolrServer.
Finally, HttpSolrServer is a single point of failure if the node you
point
Thanks Shawn, this makes a lot of sense.
With WAR going away and no mention of Solr deployment strategy (see:
https://cwiki.apache.org/confluence/display/solr/Taking+Solr+to+Production)
isn't good; there is a gab in Solr's release. It feels as if Solr 5.x was
rushed out ignoring Windows Servers
Off the cuff, it sounds like you are making a POST request to the
SearchHandler (ie: /search or /query) and the Content-TYpe you are sending
is text/xml; charset=UTF-8
In the past SearchHandler might have ignored that Content-Type, but now
that structured queries can be sent as POST data,
Look at CHANGES.txt for both Lucene and Solr, there's always an
upgrading section for each release.
Best,
Erick
On Fri, Apr 17, 2015 at 5:31 AM, Toke Eskildsen t...@statsbiblioteket.dk
wrote:
Vincenzo D'Amore v.dam...@gmail.com wrote:
I have a SolrCloud cluster with 3 server, I would like to
Hi All,
Running into an issue and wanted to see if anyone had some suggestions.
We are seeing this with both solr 4.6 and 4.10.3 code.
We are running an extremely update heavy application, with millions of writes
and deletes happening to our indexes constantly. An issue we are seeing is
that
: It looks to me that f with qq is doing phrase search, that's not what I
: want. The data in the field title is Apache Solr Release Notes
if you don't wnat phrase queries then you don't want pharse queries and
that's fine -- but it wasn't clear from any of your original emails
because you
I haven't had time to really take a look at this. But read a couple of
articles regarding the hard commit and it actually makes sense. We were
seeing tlogs in the multiple GBs during ingest. I will have some time in a
couple of weeks to come back to testing indexing. Thanks for the help.
Vy
Perfect, thank you for the information -- will have a look through those
classes.
Thank you,
Levan
--
View this message in context:
http://lucene.472066.n3.nabble.com/5-1-unique-facet-function-calcDistinct-tp4200110p4200535.html
Sent from the Solr - User mailing list archive at Nabble.com.
: df and q.op are the ones you are looking for.
: You can define them in defaults section.
specifically...
https://cwiki.apache.org/confluence/display/solr/InitParams+in+SolrConfig
:
: Ahmet
:
:
:
: On Friday, April 17, 2015 9:18 PM, Bruno Mannina bmann...@free.fr wrote:
: Dear Solr
Hi,
I need some documentation/samples on how to create a SOLR schema with nested
documents.
I have been looking online but could not find anything.
Thank you in advance,
Nick Pandrea
Hi Sujit,
Many thanks for your blog post, responding to my question, and suggesting the
alternative option ☺
I think I prefer your approach because we can supply our own Comparator. The
reason is that we need to meet some strict requirements: we can only call the
external system once to
Ah, starting to see the light ... thanks for your patience.
First, this is a Java REST service using solrj. I am using default transport
(wt=javabin, I think).
But right-clicking the URL at the top of the Admin query page and selecting
open in new tab displays the non-truncated _version_
I've posted the issue here, please let me know if any additional information
needs to be provided.
https://issues.apache.org/jira/browse/SOLR-7417
Happy to provide the feedback, using the sub-facets has been a lot of fun,
the nested facet query is especially useful.
--
View this message in
On 4/17/2015 2:15 PM, Rishi Easwaran wrote:
Running into an issue and wanted to see if anyone had some suggestions.
We are seeing this with both solr 4.6 and 4.10.3 code.
We are running an extremely update heavy application, with millions of writes
and deletes happening to our indexes
Thanks Shawn for the quick reply.
Our indexes are running on SSD, so 3 should be ok.
Any recommendation on bumping it up?
I guess will have to run optimize for entire solr cloud and see if we can
reclaim space.
Thanks,
Rishi.
-Original Message-
From: Shawn Heisey
Agreed, I also prefer the second way. I find it more readible, less verbose
while communicating the same information, less confusing to mentally parse
(is 'terms' the name of my facet, or the type of my facet?...), and less
prone to syntactlcally valid, but logically invalid inputs. Let's break
Ah, I meant SOLR-7418 https://issues.apache.org/jira/browse/SOLR-7418.
On Fri, Apr 17, 2015 at 4:30 PM, Anshum Gupta ans...@anshumgupta.net
wrote:
Hi Ere,
Those seem like valid issues. I've created an issue : SOLR-7275
https://issues.apache.org/jira/browse/SOLR-7275 and will create more as
Hello All,
I am new to solr and trying to configure highlighting. If I look at the
result in xml, or json format, I can see the highlighting part of the
data and it looks good. However the velocity page does not show the
highlighted words on my result page. Do I need to do something extra for
Hi folks,
In my DB, my records are nested in a folder base hierarchy:
Root
Level_1
record_1
record_2
Level_2
record_3
record_4
Level_3
record_5
Level_1
Level_2
Level_3
record_6
Does anyone have any thoughts on the current general structure of JSON facets?
The current general form of a facet command is:
facet_name : { facet_type : facet_args }
For example:
top_authors : { terms : {
field : author,
limit : 5,
}}
One alternative I considered in the past is having
Not unless you provide a lot more details. Specifically, anything in your Solr
logs that looks suspicious _and_ in your container logs (Tomcat? Jetty?).
Plus the message you sent.
Please review:
http://wiki.apache.org/solr/UsingMailingLists
Best,
Erick
On Thu, Apr 16, 2015 at 10:57 PM, Pavel
On 4/17/2015 7:45 PM, Vincenzo D'Amore wrote:
Hi Shawn,
thanks for your answer.
I apologise for my english, for floating results I meant random results
in queries.
As far as I know, we should split the synonyms file because of zookeeper,
there is a limit in the size of files (1MB).
All my
Hi,
As per this article, the linux machine is preferred to have 1.5 times RAM
with respect to index size. So, to verify this, I tried testing the solr
performance in different volumes of RAM allocation keeping other
configuration (i.e Solid State Drives, 8 core processor, 64-Bit) to be same
in
Hi,
This may be irrelevant but your machine configuration reminded me of some
reading I had done some time back on memory vs ssd.
Do a search on solr ssd and you should get some meaningful posts.
Like this one https://sbdevel.wordpress.com/2013/06/06/memory-is-overrated/
Regards
Puneet
On 18 Apr
Hi,
Because you went over 31-32 GB heap you lost the benefit of compressed
pointers and even though you gave the JVM more memory the GC may have had
to work harder. This is a relatively well educated guess, which you can
confirm if you run tests and look at GC counts, times, JVM heap memory pool
Hi Ha,
Yes, I think if you want to facet on the external field, the custom
component seems to be the best option IMO.
-sujit
On Fri, Apr 17, 2015 at 3:02 PM, ha.p...@arvatosystems.com wrote:
Hi Sujit,
Many thanks for your blog post, responding to my question, and suggesting
the
The other issue that would fix half of your problems is:
https://issues.apache.org/jira/browse/SOLR-7143
On Fri, Apr 17, 2015 at 4:35 PM, Anshum Gupta ans...@anshumgupta.net
wrote:
Ah, I meant SOLR-7418 https://issues.apache.org/jira/browse/SOLR-7418.
On Fri, Apr 17, 2015 at 4:30 PM, Anshum
Great!! Thank you very much.
On Fri, Apr 17, 2015 at 7:36 PM, Erick Erickson erickerick...@gmail.com
wrote:
Solr/Lucene are supposed to _always_ read one major version back. Thus
your 4.10 should be able to read indexes produced all the way back to
(and including) 3.x. Sometimes experimental
Hi Shawn,
thanks for your answer.
I apologise for my english, for floating results I meant random results
in queries.
As far as I know, we should split the synonyms file because of zookeeper,
there is a limit in the size of files (1MB).
All my synonyms are about 10MB.
I have tried again in dev
67 matches
Mail list logo