Thank you nutchsolruser and Shawn.
I've changed the clientPort to different port for each of the machine.
It is able to work for my another setup, in which I have 3 different
zookeeper folder, and each has its own configuration and all are using
zoo.cfg. For that setup I can start the 3 servers
not able to search on special characters like . ,_
my query
http://localhost:8983/solr/rna/select?q=name:UAE
Bwt=jsonfl=namerows=100
getting result UAE_bhdgsfsdbj
but for
http://localhost:8983/solr/rna/select?q=name:UAE_wt=jsonfl=namerows=100
no result found
I am using below field type
Hi,
I am trying to setup a SolrCloud cluster on top of Hadoop (HDP). The
upconfig and linkconfig commands were run successfully and the
configuration is now centrally manged in Zookeeper.
However, when I run the command to create a core, I am shown the following
permission denied error. This is
Text search means searching of text, and special characters are not... text.
Why are you using the standard tokenizer if you are not trying to search
for standard text?
Try using the white space tokenizer, which will preserve special characters.
That said, the word delimiter filter will remove
I have crawled a website using nutch.
When I try to index it with solr I get following error
org.apache.solr.common.SolrException: ERROR: [doc=http://xyz.htm]
unknown field 'metatag.keywords'
*unknown field 'metatag.keywords'*
I can not figure out where the error is as I have o not defined any
i'm using solr 4.4.
the query request works fine but when i try to add doc into solr cloud
(cloudSolrServer.request(updateRequest))
i get an error:
Exception in thread main
org.apache.solr.client.solrj.impl.CloudSolrServer$RouteException: Unknown type
19
at
Hello Shani,
Are you using SolrJ? Did you try using the same version of SolrJ (e.g v.
4.4 of SolrJ, if you're using Solr 4.4)? That's what generally worked for
me.
Kind regards,
Emre Sevinç
http://www.bigindustries.be/
On Wed, Apr 8, 2015 at 1:46 PM, Chaushu, Shani shani.chau...@intel.com
Hi Swaraj,
Thanks for the answers.
From my understanding We can index,
· Using DIH from db
· Using DIH from filesystem - this is where I am concentrating on.
o For this we can use SolrJ with Tika(solr cell) from Java layer in order to
extract the content and send the data
I just want to index only certain documents and there will not be any update
happening on the indexed document.
In our existing system we already have DIH implemented which indexes
document from sql server (As you said based on last index time). In this
case the metadata is there available in
Hi Pradeep,
AFAIK the mailing list doesn't allow attachments. I think pasting the
error should be enough
Best,
Andrea
On 04/08/2015 09:02 AM, Pradeep wrote:
We have installed solr-4.3.0 is our local but we are getting error.
Please find attachment. And help us to fix this error.
Thank You.
thanks Erick
--
View this message in context:
http://lucene.472066.n3.nabble.com/distributed-search-on-tables-tp4197456p4198285.html
Sent from the Solr - User mailing list archive at Nabble.com.
We have installed solr-4.3.0 is our local but we are getting error.
Please find attachment. And help us to fix this error.
Thank You.
Regards,
Pradeep
Hi Zheng,
I am not sure if this command *zkServer.cmd start zoo.cfg * works in
windows or not, but in zkServer.cmd it calls zkEnv.cmd where
*ZOOCFG=%ZOOCFGDIR%\zoo.cfg* is set. So, if you want to run multiple
instances of zookeeper, change zoo.cfg to your config file and start
zookeeper.
The
Hi Sangeetha,
/update/extract refers to extractrequesthandler.
If you only want to index the data, you can do it with extractrequesthandler.
I dont think it requires metadata, but you need to provide literal.id to
specify which field will be unique id.
For more information :-
To be precise: create one zoo.cfg for each of the instances. One config
file for all is a bad idea.
In each config file, use the same server.X lines, but use a unique
clientPort.
As you will also have separate data directories, I would recommend
having one root directory .../zookeeper where you
Hi,
Brief:
I am new to Solr and E commerce web apps and Java.
i want to integrate solr in eCommerce web application (developed using Java
on Linux).
I have following queries.
1. how to setup SolrCloud on Tomcat. Searched on internet but could not get
clear steps yet. Also tried some steps but
Hi,
I have installed Curl on debian linux. But when i use curl to create
collection i am getting http autentication error.
--
Thanks Regards,
Jainam Vora
How much RAM do you have? Check whether your system is compute-bound or
I/O-bound? If all or most of your index doesn't fit in the system memory
available for file caching, you're asking for trouble.
Is the indexing time also unacceptably slow, or just the query time?
-- Jack Krupansky
On Wed,
Hi,
I tried to get the SolrJ in older version, but I'm using solr-spark package and
it fails with compilation errors probably because it uses function from newer
versions...
I can't find any solution...
-Original Message-
From: Emre Sevinc [mailto:emre.sev...@gmail.com]
Sent:
On 4/8/2015 6:30 AM, Chaushu, Shani wrote:
I tried to get the SolrJ in older version, but I'm using solr-spark package
and it fails with compilation errors probably because it uses function from
newer versions...
I can't find any solution...
Looking at the github repo for spark-solr, I see
Hi,
I have a core with about 20M documents and the size on disc is about
50GB. It is running on a single EC2 instance. If the core is warmed up,
everything is running fine. The problem is the following:
We assign categories (similar to tags) to documents. Those are stored in
a multivalue string
Thank you Swaraj and Jurgen for the information.
I'll just stick to the one zoo.cfg for each instance. Now I have one root
directory .../zookeeper where I create the 3 subdirectories for each of the
instance (known as zookeeper1, zookeeper2 and zookeeper3), and each of them
have their own
Hello All,
I am using Master - Slave architecture setup with hundreds of cores getting
replicated between master and slave servers. I am facing very weird issue
while creating a new core.
Whenever there is a new call for a new core creation (using
Well, at least it's _some_ progress ;).
Agreed, the segments hanging around is still something of a mystery
although if I really stretch I could relate them, maybe.
I believe there's clean-up logic when a core starts up to nuke cruft
in the index directory. If the cruft was created after a core
bq: After the commit, query times are unacceptable slow
First, please quantify unacceptable. 100ms? 10,000ms? Details matter.
Second, the purpose of autowarming is exactly to smooth out the first few
searches when a new searcher is opened, are you doing any?
Third: What are your autocommit
http://localhost:8983/solr/rna/select?q=test_name:*Uae
blow*wt=jsonrows=100
getting
{
responseHeader: {
status: 400,
QTime: 28
},
error: {
msg: no field name specified in query and no default specified via 'df'
param,
code: 400
}
}
plz help!!
--
View this message in context:
Re,
You have to specify defautSearchField tag in Schema.xml
Regards,Andy
Le Mercredi 8 avril 2015 21h33, avinash09 avinash.i...@gmail.com a écrit
:
http://localhost:8983/solr/rna/select?q=test_name:*Uae
blow*wt=jsonrows=100
getting
{
responseHeader: {
status: 400,
QTime: 28
},
See inline for a few answers:
On Wed, Apr 8, 2015 at 10:32 AM, jainam vora jainam.v...@gmail.com wrote:
Hi,
Brief:
I am new to Solr and E commerce web apps and Java.
i want to integrate solr in eCommerce web application (developed using Java
on Linux).
I have following queries.
1. how
Dear all,
Hi,
As a part of my code I have to update Lucene document. For this purpose I
used writer.updateDocument() method. My problem is the update process is
not affect index until restarting Solr. Would you please tell me what part
of my code is wrong? Or what should I add in order to apply
Specifying the default search field in the schema has been
deprecated for a while, it wasn't flexible enough.
The recommended way is to specify a df parameter in
your request handler defaults.
The space separates the field specification from the second term.\
Assuming that you want to search
Wups - sorry folks, I send this prematurely. After typing this out I think
I have it figured out - although SPLITSHARD ignores maxShardsPerNode,
ADDREPLICA does not. So ADDREPLICA fails because I already have too many
shards on a single node.
On Wed, Apr 8, 2015 at 11:18 PM, Ian Rose
Thanks, I figured that might be the case (hand-editting clusterstate.json).
- Ian
On Wed, Apr 8, 2015 at 11:46 PM, ralph tice ralph.t...@gmail.com wrote:
It looks like there's a patch available:
https://issues.apache.org/jira/browse/SOLR-5132
Currently the only way without that patch is to
Found this error which likely explains my issue with new replicas not coming
up, not sure next step. Almost looks like Zookeeper's record of a Shard's
leader is not being updated?
4/8/2015, 4:56:03 PM
ERROR
ShardLeaderElectionContext
There was a problem trying to register as the
Hello,
When creating a new replica, and the state is recorded as active with in ZK
clusterstate, does that mean that new replica has synched with the leader
replica for the particular shard?
Thanks,
Matt
I am extremely new to SOLR and am wondering if it is possible to do something
like the following. Basically I have been tasked with researching SOLR to see
if we can replace our current searching algorithm.
We have a website with product data. Product data includes standard things
like Name,
On my local machine I have the following test setup:
* 2 nodes (JVMs)
* 1 collection named testdrive, that was originally created with
numShards=1 and maxShardsPerNode=1.
* After a series of SPLITSHARD commands, I now have 4 shards, as follows:
testdrive_shard1_0_0_replica1 (L) Active 115
Hey guys, it seems that omitTermFreqAndPositions is not very usable with
edismax, and I'm wondering if this is intended behavior, and how I can get
around the problem.
The setup:
define field foo with omitTermFreqAndPositions=true
The query:
q=ground coffeeqf=foo bar baz
The error:
It looks like there's a patch available:
https://issues.apache.org/jira/browse/SOLR-5132
Currently the only way without that patch is to hand-edit
clusterstate.json, which is very ill advised. If you absolutely must,
it's best to stop all your Solr nodes, backup the current clusterstate
in ZK,
I previously created several collections with maxShardsPerNode=1 but I
would now like to change that (to unlimited if that is an option). Is
changing this value possible?
Cheers,
- Ian
Ryan:
bq: I don't want it to issue phrase queries to that field ever
This is one of those requirements that you'd have to enforce at the
app layer. Having Solr (or Lucene) enforce a rule like this for
everyone would be terrible.
So if you're turning off TF but also saying title is one of the
Matt:
In a word, yes. Depending on the size of the index for that shard,
the transition from Down-Recovering-Active may be too fast to catch.
If replicating the index takes a while, though, you should at least
see the Recovering state, during which time there won't be any
searches forwarded to
I have a solr cloud instance with 8 nodes, 4 shards and facing memory leak on
the JVMs
here are the details of the instance.
1. 8 nodes, 4 shards(2 nodes per shard)
2. each node having about 55 GB of Data, in total there is 450 million
documents in the collection. so the document size is not
Thanks for your thought Shawn, I don't think fq will be helpful here. The
field for which I want to turn TF off is title, which is actually one of
the primary components of score, so I really need it in qf. I just don't
want the TF portion of the score for that field only. I don't want it to
Hi Matt,
If it's the replica state that you're looking at, yes, it means that the
Replica is in sync with the leader and serving/ready to serve requests.
On Wed, Apr 8, 2015 at 2:58 PM, Matt Kuiper matt.kui...@issinc.com wrote:
Hello,
When creating a new replica, and the state is recorded as
Erick, Anshum,
Thanks for your replies! Yes, it is replica state that I am looking at, and
this the answer I was hoping for.
I am working on a solution that involves moving some replicas to new Solr nodes
as they are made available. Before deleting the original replicas backing the
shard,
On 4/8/2015 5:06 PM, Ryan Josal wrote:
The error:
IllegalStateException: field foo indexed without position data; cannot
run PhraseQuery.
It would actually be ok for us to index position data but there isn't an
option for that without term frequencies. No TF is important for us when
it
Matt:
How are you creating the new replica? Are you giving it an explicit
name? And especially is it the same name as one you've already
deleted?
'cause I can't really imagine why you'd be getting a ZK exception
saying the node already exists.
Shot in the dark here..
On Wed, Apr 8, 2015 at
Hi All,
We are trying to setup the Solr Cloud in our team and able setup multiple nodes
in one server as a cloud.
Need clarifications on the following.
Is there any good documentation, which can help us to build the Solr Cloud with
multiple physical servers?
Since the Solr Cloud is
Are there at least a small number of categories of users with discrete
prices, or can each user have their own price. The former is doable, the
latter is not unless the number of users is relatively small, in which case
they are equivalent to categories.
You could have a set of dynamic fields,
49 matches
Mail list logo