Hi,
Issue in brief:
I am facing a strange issue, where, the collections that are deleted in
SOLR, are still having reference in Zookeeper and due to which, in the solr
cloud console, i am still seeing the reference to the deleted collections in
down state
Issue in Detail:
I am using Solr 4.5.1
Hi Radha,
This is strange as I the collections API delete command is supposed to
clean up zk. Do you see any errors in your Solr logs? Does the response
from the call include any errors/exceptions?
On Sun, Sep 7, 2014 at 11:32 PM, RadhaJayalakshmi
rlakshminaraya...@inautix.co.in wrote:
Hi,
Hi Susheel ,
Thanks for the information.
I have crawled few website and all I need is for sentence tokenizers on the
data I have collected.
These websites are English only.
Well I don't have experience in writing custom sentence tokenizers for
solr. Is there any tutorial link which tell how to do
Hi Shawn,
I try to fix the problem with original jars and now just discovered it
happens only with SolrJ client when doing a soft commit with the
following method call signature:
solr.commit(false, false, true);
Where solr is an HttpSolrServer initialized at the beginning of the
Is there any open source or commercial analyzer for Kannada language?
If anyone have experience with indexing Kannada documents, please share the
relevant information.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-Analyzer-for-Kannada-language-tp4157382.html
Sent
On 09/08/2014 02:55 AM, Erick Erickson wrote:
I really recommend you use the new-style core discovery, if for no
other reason than this style is deprecated in 5.0. See:
https://wiki.apache.org/solr/Solr.xml%204.4%20and%20beyond
Oh I didn't know that.
Anyway problem I experienced was result of
Out of curiosity, what would be an example query for your application that
would depend on sentence tokenization, as opposed to simple term
tokenization? I mean, there are no sentence-based query operators in the
Solr query parsers.
-- Jack Krupansky
-Original Message-
From: Sandeep
You also need to take a stance as to whether you wish to auto-detect the
language at query time vs. have a UI selection of language vs. attempt to
perform the same query for each available language and then determine
which has the best relevancy. The latter two options are very sensitive to
Some further details out of my mind:
- it is a stream based feature
- IDF estimates get updated and refined as more and more documents pass through
- it is actually IDF weighting with stopwords and boosting
-- stopwords should be ignored and not get vectorized
-- boosting should give some boost to
In one of the talks by Trey Grainger (author of Solr in Action) it touches how
on CareerBuilder are dealing with multilingual with payloads, its a little more
of work but I think it would payoff.
On Sep 8, 2014, at 7:58 AM, Jack Krupansky j...@basetechnology.com wrote:
You also need to take
Whew! Thanks for letting us know...
Yeah, the core discovery mode seems easier to
use when maintaining lots of cores since solr.xml
can be a pain.
That said, whatever works of course.
Erick
On Mon, Sep 8, 2014 at 4:28 AM, Jakov Sosic jso...@gmail.com wrote:
On 09/08/2014 02:55 AM, Erick
I had a similar need. The resulting tool is in scala, but it still might
be useful to look at. I had to work through some of those same issues:
https://github.com/whitepages/solrcloud_manager
From a clusterstate perspective, I mostly cared about active vs
non-active, so here¹s a sample output
Joel and Erick,
Thank you very much for explaining how the ReRanking works. Now
its a bit more clear.
Thanks,
Ravi Kiran Bhaskar
On Sun, Sep 7, 2014 at 4:45 PM, Joel Bernstein joels...@gmail.com wrote:
Oops wrong usage pattern. It should be:
1) Main query is sorted by a field
: Is there a list of possible other statuscodes you can receive in case
: anything fails and what these errorcodes mean?
:
: I don't think we have a list of possible other status because Solr
: doen't return status other than 0. Instead of status code in XML,
: you should look at HTTP status
I just got Solr 4.9.0 running as a 3 node cloud. I use the CloudSolrServer
class to connect and do queries, but it isn't working now using HTTPS. I
don't see any options for the CloudSolrServer to use https (no key/trust
store or anything).
What SolrJ classes should I be looking at to connect
I am trying to setup 2 shard cluster with 2 replicas with dedicated nodes
for replicas. I have 4 node SolrCloud setup that I am trying to shard
using collections api .. (Like
https://wiki.apache.org/solr/SolrCloud#Example_C:_Two_shard_cluster_with_shard_replicas_and_zookeeper_ensemble
)
I ran
Hi Chris,
Check out the Solr Reference Guide SolrJ example indexing a doc over HTTPS
using CloudSolrServer:
https://cwiki.apache.org/confluence/display/solr/Enabling+SSL#EnablingSSL-IndexadocumentusingCloudSolrServer
Steve
www.lucidworks.com
On Sep 8, 2014, at 2:19 PM, Christopher Gross
Ahhh, this is a continual source of confusion. I've started a one-man
campaign to talk about leaders and followers when relevant...
_Every_ node is a replica. This is because a node can be a leader or
follower, and the role can change.
So your case is entirely normal. These nodes are probably
On Mon, Sep 8, 2014 at 3:11 PM, Erick Erickson erickerick...@gmail.com wrote:
I've started a one-man
campaign to talk about leaders and followers when relevant
Well, if you write it up on the Wiki/Manual and keep pointing people
to it, maybe we will all fall in line. I, for one, do not care
Thanks Erick. That cleared my confusion.
I have a follow up question - If I run the CREATE command with 4 nodes in
createNodeSet, I thought 2 leaders and 2 followers will be created
automatically. Thats not the case, however.
You need to specify a replication factor of 2 if you want two copies of
each shard. Solr doesn¹t ³auto fill² available capacity, contrary to the
misleading examples on the http://wiki.apache.org/solr/SolrCloud page.
Those examples only have that behavior because they ask you to copy the
examples
Hi Ilia,
When writing *Solr in Action*, I implemented a feature which can do what
you're asking (allow multiple, dynamic analyzers to be used in a single
text field). This would allow you to use the same field and dynamically
change the analyzers (for example, you could do language-identification
Sandeep,
As Jack mentioned it will be useful to know the use case/what kind of query you
will be executing as you may also need to handle on query side not just on
indexing side. For integrating with nltk there could be different options like
calling ntlk as out of proc or use jythonc to
There would be records in millions and field would only be md5 hash. Size
of field would be 32 chars
Can you please point out what would be the best approach ?
Thanks
On Sep 5, 2014 12:47 PM, Mikhail Khludnev mkhlud...@griddynamics.com
wrote:
Hello,
You are asking without giving a context.
We have a redundant data center in case the primary goes down. Currently we
have 1 master and multiple slaves on primary data center. This master also
replicates to a slave in secondary data center. So if the primary goes down
at least the read only part works. However, now we want writes to work
25 matches
Mail list logo