On 12/22/2011 4:39 AM, Dean Pullen wrote:
Yeh the drop index via the URL command doesn't help anyway - when rebuilding
the index the timestamp is obviously ahead of master (as the slave is being
created now) so the replication will still not happen.
If you deleted the index and create the
On 1/6/2012 9:57 AM, Phillip Rhodes wrote:
On Fri, Jan 6, 2012 at 3:06 AM, Sven Maurmann s...@kippdata.de wrote:
Hi,
from your snippets the reason is not completely clear. There are a number of
reasons for not starting up the
server. For example in case of a faulty configuration of the core
On 1/6/2012 10:19 AM, Phillip Rhodes wrote:
2012/1/6 Yury Kats yuryk...@yahoo.com:
Have you tried passing core name (collection1) to the c'tor, instead
of the empty string?
Yep, but that gives the same error (with the core name appended) such
as no such core: collection1
That probably
I'm running into a problem with queries that contain forward slashes and more
than one field.
For example, these queries work fine:
fieldName:/a
fieldName:/*
But if I have two fields with similar syntax in the same query, it fails.
For simplicity, I'm using the same field twice:
fieldName:/a
On 2/22/2012 12:25 PM, Yury Kats wrote:
I'm running into a problem with queries that contain forward slashes and more
than one field.
For example, these queries work fine:
fieldName:/a
fieldName:/*
But if I have two fields with similar syntax in the same query, it fails
On 2/22/2012 1:05 PM, Em wrote:
Yury,
are you sure your request has a proper url-encoding?
Yes
On 2/22/2012 1:25 PM, Em wrote:
That's strange.
Could you provide a sample dataset?
Data set does not matter. The query fails to parse, long before it gets to the
data.
On 2/22/2012 1:24 PM, Yonik Seeley wrote:
This is a bit puzzling as the forward slash is not part of the query
language, is it?
Regex queries were added that use forward slashes:
https://issues.apache.org/jira/browse/LUCENE-2604
Oh, so / is a special character now? I don't think it is
On 2/22/2012 1:24 PM, Yonik Seeley wrote:
Looks like escaping forward slashes makes the query work, eg
fieldName:\/a fieldName:\/a
This is a bit puzzling as the forward slash is not part of the query
language, is it?
Regex queries were added that use forward slashes:
On 5/18/2012 9:54 AM, Tolga wrote:
Hi,
I've put the line copyField=* dest=text stored=true
indexed=true/ in my schema.xml and restarted Solr, crawled my
website, and indexed (I've also committed but do I really have to
commit?). But I still have to search with content:mykeyword at the
On 5/18/2012 4:02 PM, Tolga wrote:
Default field? I'm not sure but I think I do. Will have to look.
http://wiki.apache.org/solr/SchemaXml#The_Default_Search_Field
On 12/30/2012 11:57 AM, uwe72 wrote:
Hi there,
how can i add a date field to a pdf document?
Same way you add the ID field, using literal parameter.
ContentStreamUpdateRequest up = new
ContentStreamUpdateRequest(/update/extract);
up.addFile(pdfFile, application/octet-stream);
On 12/30/2012 3:55 PM, uwe72 wrote:
but i can just add String values.i want to add Date objects?!
You represent the Date as a String, in format Solr uses for dates:
http://lucene.apache.org/solr/api-4_0_0-BETA/org/apache/solr/schema/DateField.html
On 1/9/2013 10:38 AM, Shahar Davidson wrote:
Hi All,
I have a client app that uses SolrJ and which requires to collect the names
(and just the names) of all loaded cores.
I have about 380 Solr Cores on a single Solr server (net indices size is
about 220GB).
Running the STATUS action
When crating cores through solr.xml, I am able to specify custom
properties, to be referenced in solrconfig.xml. For example:
cores adminPath=/admin/cores defaultCoreName=master
core name=master instanceDir=core1 shard=shard1 collection=myconf
property name=enable.master value=true /
I have multiple SolrCloud instances, each running its own Zookeeper
(Solr launched with -DzkRun).
I would like to create an ensemble out of them. I know about -DzkHost
parameter, but can I achieve the same programmatically? Either with
SolrJ or REST API?
Thanks,
Yury
On 8/8/2011 12:00 AM, Daniel Schobel wrote:
Can someone provide me with a succinct defintion of what a solr core
is? Is there a one-to-one relationship of cores to solr indices or can
you have multiple indices per core?
http://wiki.apache.org/solr/CoreAdmin
There's one index per core.
On 8/8/2011 4:34 PM, Jason Toy wrote:
Aelexei, thank you , that does seem to work.
My sort results seem to be totally wrong though, I'm not sure if its because
of my sort function or something else.
My query consists of:
sort=termfreq(all_lists_text,'indie+music')+descq=*:*rows=100
And I
On 8/8/2011 5:03 PM, Matt Shields wrote:
I'm looking for some examples of how to setup Solr on EC2. The
configuration I'm looking for would have multiple nodes for redundancy.
I've tested in-house with a single master and slave with replication
running in Tomcat on Windows Server 2003, but
On 9/7/2011 3:18 PM, Pulkit Singhal wrote:
Hello,
I'm working off the trunk and the following wiki link:
http://wiki.apache.org/solr/SolrCloud
The wiki link has a section that seeks to quickly familiarize a user
with replication in SolrCloud - Example B: Simple two shard cluster
with
On 9/9/2011 10:52 AM, Pulkit Singhal wrote:
Thank You Yury. After looking at your thread, there's something I must
clarify: Is solr.xml not uploaded and held in ZooKeeper?
Not as far as I understand. Cores are loaded/created by the local
Solr server based on solr.xml and then registered with
On 9/9/2011 4:48 PM, Jamie Johnson wrote:
When doing writes do all writes need to be done to the primary shard
or are writes that are done to the replica also pushed to all replicas
of that shard?
If you have replication setup between cores, all changes to the
slave will be overwritten by
On 9/9/2011 6:54 PM, Pulkit Singhal wrote:
Thanks Again.
Another question:
My solr.xml has:
cores adminPath=/admin/cores defaultCoreName=master1
core name=master1 instanceDir=. shard=shard1 collection=myconf/
/cores
And I omitted -Dcollection.configName=myconf from the startup
On 9/10/2011 3:54 PM, Pulkit Singhal wrote:
Hi Yury,
How do you manage to start the instances without any issues? The way I see
it, no matter which instance is started first, the slave will complain about
not being to find its respective master because that instance hasn't been
started yet
On 9/11/2011 11:24 PM, William Bell wrote:
I am using 3.3 SOLR. I tried passing in -Denable.master=true and
-Denable.slave=true on the Slave machine.
Then I changed solrconfig.xml to reference each as per:
One of my users observed that the index size (in bytes)
increased over night. There was no indexing activity
at that time, only querying was taking place.
Running optimize brought the index size back down to
what it was when indexing finished the day before.
What could explain that?
On 9/14/2011 2:36 PM, Erick Erickson wrote:
What is the machine used for? Was your user looking at
a master? Slave? Something used for both?
Stand-alone machine with multiple Solr cores. No replication.
Measuring the size of all the files in the index? Or looking
at memory?
Disk space.
Let's say I'm forcing a replication of a core using fetchindex command.
No new content is being added to the master.
I can check whether replication has finished by periodically querying
master and slave for their indexversion and comparing the two.
But what's the best way to check if
On 9/16/2011 4:58 PM, Brandon Fish wrote:
Hi Yury,
You could try checking out the details command of the replication handler:
http://slave_host:port/solr/replication?command=details
which has information such as isReplicating.
How reliable is isReplicating? Is it updated on unexpected
On 9/23/2011 6:00 PM, hadi wrote:
I index my files with solrj and crawl my sites with nutch 1.3 ,as you
know, i have to overwrite the nutch schema on solr schema in order to
have view the result in solr/browse, in this case i should define two
cores,but i want have single result or the user
On 9/24/2011 3:09 AM, hadi wrote:
I do not know how to search both cores and not define shard
parameter,could you show me some solutions for solve my issue?
See this: http://wiki.apache.org/solr/DistributedSearch
On 9/27/2011 5:16 PM, Darren Govoni wrote:
On 09/27/2011 05:05 PM, Yury Kats wrote:
You need to either submit the docs to both nodes, or have a replication
setup between the two. Otherwise they are not in sync.
I hope that's not the case. :/ My understanding (or hope maybe) is that
the new
Nope
On 9/29/2011 12:17 AM, Pulkit Singhal wrote:
Did you find out about this?
2011/8/2 Yury Kats yuryk...@yahoo.com:
I have multiple SolrCloud instances, each running its own Zookeeper
(Solr launched with -DzkRun).
I would like to create an ensemble out of them. I know about -DzkHost
On 9/29/2011 7:22 AM, Darren Govoni wrote:
That was kinda my point. The new cloud implementation
is not about replication, nor should it be. But rather about
horizontal scalability where nodes manage different parts
of a unified index.
It;s about many things. You stated one, but there are
On 9/30/2011 12:26 PM, Pulkit Singhal wrote:
SOLR-2355 is definitely a step in the right direction but something I
would like to get clarified:
Questions about SOLR-2355 are best asked in SOLR-2355 :)
b) Does this basic implementation distribute across shards or across
cores?
From a brief
On 10/14/2011 9:29 AM, Rohit wrote:
I want to user POST instead of GET while using solrj, but I am unable to
find a clear example for it. If anyone has implemented the same it would be
nice to get some insight.
To do what? Submit? Query? How do you use SolrJ now?
On 10/14/2011 12:11 PM, Rohit wrote:
I want to query, right now I use it in the following way,
CommonsHttpSolrServer server = new CommonsHttpSolrServer(URL HERE);
SolrQuery sq = new SolrQuery();
sq.add(q,query);
QueryResponse qr = server.query(sq);
QueryResponse qr = server.query(sq,
On 10/19/2011 5:15 PM, Darren Govoni wrote:
Hi Otis,
Yeah, I saw page, but it says for merging cores, which I presume
must reside locally to the solr instance doing the merging?
What I'm interested in doing is merging across solr instances running on
different machines into a single
On 10/20/2011 11:33 AM, Rahul Warawdekar wrote:
Hi,
I am trying to evaluate distributed search for my project by splitting up
our single index on 2 shards with Solr 3.1
When I query the first solr server by passing the shards parameter, I get
correct search results from both shards.
(
On 10/25/2011 11:24 AM, Mark Schoy wrote:
Hi,
is ist possible to define a relative path in confFile?
For example:
str name=confFiles../../x.xml/str
If yes, to which location will the file be copied at the slave?
I don;t think it's possible. Replication copies confFiles from master
There's a defaultCore parameter in solr.xml that let's you specify what core
should be used when none is specified in the URL. You can change that every
time you create a new core.
From: Vadim Kisselmann v.kisselm...@googlemail.com
To:
On 11/3/2011 12:59 PM, Milan Dobrota wrote:
Is there any way to define the default value for the dynamic fields in
SOLR? I use some dynamic fields of type float with _val_ and if they
haven't been created at index time, the value defaults to 0. I would want
this to be 1. Can that be changed?
On 11/12/2011 4:08 PM, mikr00 wrote:
Similar to a first in first out list. The problem is: It's easy to check the
limit, but how can I delete the oldest documents to go again below the
limit? Can I do it with a delete by query request? In that case, I would
probably have to limit the number of
On 12/9/2011 11:54 PM, Rohit wrote:
Hi All,
Don't know if this question is directly related to this forum, I am running
Solr in Tomcat on linux server. The moment I start tomcat the virtual memory
shown using TOP command goes to its max 31.1G and then remains there.
Is this the
On 12/11/2011 4:57 AM, Rohit wrote:
What are the difference in the different DirectoryFactory?
http://lucene.apache.org/java/3_3_0/api/core/org/apache/lucene/store/MMapDirectory.html
http://lucene.apache.org/java/3_3_0/api/core/org/apache/lucene/store/NIOFSDirectory.html
Solr server with a Luke request,
eg http://localhost:8983/solr/admin/luke
Dmitry
2011/12/12 Yury Kats yuryk...@yahoo.com
On 12/11/2011 4:57 AM, Rohit wrote:
What are the difference in the different DirectoryFactory?
http://lucene.apache.org/java/3_3_0/api/core/org/apache/lucene/store
Does anybody have an idea, or better yet, measured data,
to see what the overhead of a core is, both in memory and speed?
For example, what would be the difference between having 1 core
with 100M documents versus having 10 cores with 10M documents?
On 12/15/2011 1:07 PM, Robert Stewart wrote:
I think overall memory usage would be close to the same.
Is this really so? I suspect that the consumed memory is in direct
proportion to the number of terms in the index. I also suspect that
if I divided 1 core with N terms into 10 smaller cores,
On 12/15/2011 1:41 PM, Robert Petersen wrote:
loading. Try it out, but make sure that the functionality you are
actually looking for isn't sharding instead of multiple cores...
Yes, but the way to achieve sharding is to have multiple cores.
The question is then becomes -- how many cores
On 12/15/2011 4:46 PM, Robert Petersen wrote:
Sure that is possible, but doesn't that defeat the purpose of sharding?
Why distribute across one machine? Just keep all in one index in that
case is my thought there...
To be able to scale w/o re-indexing. Also often referred to as
Hi,
I have two Solr nodes, each managing two cores -- a master core and a slave
core.
The slaves are setup to replicate from the other node's masters
That is, node1.master - node2.slave, node2.master - node1.slave.
The replication is configured in each core's solrconfig.xml, eg
Master's
On 5/17/2011 10:17 AM, Stefan Matheis wrote:
Yury,
perhaps Java-Pararms (like used for this sample:
http://wiki.apache.org/solr/SolrReplication#enable.2BAC8-disable_master.2BAC8-slave_in_a_node)
can help you?
Ah, thanks! It does seem to work!
Cluster's solrconfig.xml (shared between all
I would create a replication slave, for which you can specify whatever
location you want, even put it on a different machine. If ran on the same
machine, the slave can be another core in the same Solr instance.
On 5/17/2011 2:20 PM, Dietrich wrote:
I am using Solr Replication to create a
On 5/18/2011 4:19 PM, Judioo wrote:
Any help is greatly appreciated. Pointers to documentation that address my
issues is even more helpful.
I think this would be a good start:
http://wiki.apache.org/solr/DataImportHandler#Usage_with_XML.2BAC8-HTTP_Datasource
On 5/25/2011 9:29 AM, Romi wrote:
and in http://wiki.apache.org/solr/SchemaXml#Fields it is clearly mentioned
that a non-indexed field is not searchable then why i am getting search
result. why should stored=true matter if indexed=false
indexed controls whether you can find the document based
On 5/27/2011 6:48 AM, Romi wrote:
What is the benifit of setting autocommit in solrconfig.xml.
i read somewhere that these settings control how often pending updates will
be automatically pushed to the index.
does it mean if solr server is running then it automaticaly starts indexing
process
On 6/14/2012 2:05 AM, Daniel Brügge wrote:
Will check later to use different data dirs for the core on
each instance.
But because each Solr sits in it's own openvz instance (virtual
server respectively) they should be totally separated. At least
from my point of understanding virtualization.
On 6/15/2012 12:49 PM, Otis Gospodnetic wrote:
Hi,
How exactly does SolrCloud handle split brain situations?
Imagine a cluster of 10 nodes.
Imagine 3 of them being connected to the network by some switch and imagine
the out port of this switch dies.
When that happens, these 3 nodes will
On 6/19/2012 4:06 PM, Justin Babuscio wrote:
Solr v3.5.0
8 Master Shards
2 Slaves Per Master
Confirming that there are no active records being written, the numFound
value is decreasing as we page through the results.
For example,
Page1 - numFound = 3683
Page2 - numFound = 3683
Page3 -
On 6/25/2012 8:40 AM, Yuval Dotan wrote:
Hi
Is there a *programmatic (java) *way to connect to the Solr server (using
solrj probably) and get the number of cores and core names?
A STATUS admin request will give you all available cores, with their names.
Sorting is a memory-intensive operation indeed.
Not sure what you are asking, but it may very well be that your
only option is to give JVM more memory.
On 7/10/2012 8:25 AM, Bruno Mannina wrote:
Dear Solr Users,
Each time I try to do a request with sort=pubdate+desc
I get:
GRAVE:
but the JVM accepts only 2Go max with the option
-Xmx
is it normal?
Thanks,
Bruno
Le 11/07/2012 03:42, Yury Kats a écrit :
Sorting is a memory-intensive operation indeed.
Not sure what you are asking, but it may very well be that your
only option is to give JVM more memory.
On 7/10/2012 8:25
On 7/11/2012 2:55 PM, Alexander Aristov wrote:
content:?? doesn't work :)
I would try escaping them: content:\?\?\?\?\?\?
I have an indexed, not stored, not multiValued field in the schema.
If I change this field to be multiValued, would I need to re-index
everything, or would all existing documents (that were indexed while
the field was not multiValued) still be queryable?
Thanks,
Yury
On 7/17/2012 9:26 PM, Zhang, Lisheng wrote:
Thanks very much for quick help! Multicore sounds interesting,
I roughly read the doc, so we need to put each core name into
Solr config XML, if we add another core and change XML, do we
need to restart Solr?
You can add/create cores on the fly,
On 7/18/2012 7:11 PM, Briggs Thompson wrote:
I have realized this is not specific to SolrJ but to my instance of Solr.
Using curl to delete by query is not working either.
Can be this: https://issues.apache.org/jira/browse/SOLR-3432
On 8/16/2012 6:57 AM, Muzaffer Tolga Özses wrote:
Also, below are the lines I got when starting it:
SEVERE: org.apache.solr.common.SolrException: Schema Parsing Failed:
multiple points
...
Caused by: java.lang.NumberFormatException: multiple points
at
I'm sending streams of data to Solr, using ExtractingRequestHandler to be
parsed/extracted by Tika and then indexed.
While multiple streams can be passed with a single request to Solr, each stream
ends up being indexed into a separate document.
Or, if I pass the unique id parameter with the
Answering my own question, for archive's sake,
I worked this out by creating my own UpdateRequestProcessor.
On 10/4/2012 2:35 PM, Yury Kats wrote:
I'm sending streams of data to Solr, using ExtractingRequestHandler to be
parsed/extracted by Tika and then indexed.
While multiple streams can
You can merge indexes. You cannot split them.
jefferyyuan yuanyun...@gmail.com wrote:
Thanks for the reply, but I think SolrReplication may not help in this case,
as we don't want to replicate all indexs to solr2, just a part of
index(index of doc created by me). Seems SolrReplication doesn't
I'm pretty sure this problem has been there forever -- the parsing of zkHost is
busted. I believe it's only been intended for example/demo purposes and
therefore makes some assumptions about the value.
I haven't looked at the current code, but this is my recollection from about a
year ago.
No, the implementation was very specific to my needs.
On 5/27/2013 8:28 AM, Alexandre Rafalovitch wrote:
You did not open source it by any chance? :-)
Regards,
Alex.
I have seen this with very few indexed documents and multiple shards.
In such a case, some shards may not have any documents, and when the query
happens to hit such a shard, it does not find the fields it's looking for
and turns this into "column not found". If you resubmit the query and hit
a
73 matches
Mail list logo