il.com> wrote:
> On Mon, Apr 18, 2016 at 3:52 PM, John Bickerstaff
> <j...@johnbickerstaff.com> wrote:
> > Thanks all - very helpful.
> >
> > @Shawn - your reply implies that even if I'm hitting the URL for a single
> > endpoint via HTTP - the "balancing&qu
fine
> -
> > although the search requests may not run on the Solr instance the load
> > balancer targeted - due to "a" above.
> >
> > Corrections or refinements welcomed...
> >
> > On Mon, Apr 18, 2016 at 7:21 AM, Shawn Heisey <apa...@elyograg
un on the Solr instance the load
balancer targeted - due to "a" above.
Corrections or refinements welcomed...
On Mon, Apr 18, 2016 at 7:21 AM, Shawn Heisey <apa...@elyograg.org> wrote:
> On 4/17/2016 10:35 PM, John Bickerstaff wrote:
> > My prior use of SOLR in produ
hould still be distributed with a load balancer.
Queries do NOT go through Zookeeper.
wunder
Walter Underwood
wun...@wunderwood.org
http://observer.wunderwood.org/ (my blog)
> On Apr 17, 2016, at 9:35 PM, John Bickerstaff <j...@johnbickerstaff.com>
wrote:
>
> My prior use of SOLR
My prior use of SOLR in production was pre SOLR cloud. We put a
round-robin load balancer in front of replicas for searching.
Do I understand correctly that a load balancer is unnecessary with SOLR
Cloud? I. E. -- SOLR and Zookeeper will balance the load, regardless of
which replica's URL is
Oh, and what, if any directories need to exist for the ADDREPLICA command
to work?
Hopefully nothing past the already existing /var/solr/data created by the
Solr install script?
On Fri, Apr 15, 2016 at 11:18 AM, John Bickerstaff <j...@johnbickerstaff.com
> wrote:
> Oh, and wha
Oh, and what, if any directories need to exist for the ADDREPLICA
On Fri, Apr 15, 2016 at 11:09 AM, John Bickerstaff <j...@johnbickerstaff.com
> wrote:
> Thanks again Eric - I'm going to be trying the ADDREPLICA again today or
> Monday. I much prefer that to hand-edit hackery...
>
es the `=...` actually work for you? When attempting similar with
> > Solr 5.3.1, despite what documentation said, I had to use
> > `node_name=...`.
> >
> >
> > Thanks,
> > Jarek
> >
> > On Fri, 15 Apr 2016, at 05:48, John Bickerstaff wrote:
> >
ICA=test2=shard1=x.x.x.x:9001_solr>
(Note the / instead of _ )
On Thu, Apr 14, 2016 at 10:45 PM, John Bickerstaff <j...@johnbickerstaff.com
> wrote:
> Jay - it's probably too simple, but the error says "not currently active"
> which could, of course, mean that
u're using ports different from the default 8983 for your
Solr instances...
You probably checked already, but I thought I'd mention it.
On Thu, Apr 14, 2016 at 8:30 PM, John Bickerstaff <j...@johnbickerstaff.com>
wrote:
> Thanks Eric!
>
> I'll look into that immediately - yes, I thin
instance must be up and running for the replica to
> be added, but that's not onerous
>
>
> The bin/solr script is a "work in progress", and doesn't have direct
> support
> for "addreplica", but it could be added.
>
> Best,
> Erick
>
> On Thu, Ap
llection, and then
> inspecting the live_nodes list in Zookeeper to confirm that the (live) node
> list is actually what you think it is.
>
>
>
>
>
> On 4/14/16, 4:04 PM, "John Bickerstaff" <j...@johnbickerstaff.com> wrote:
>
> >5.4
> >
> >
5.4
This problem drove me insane for about a month...
I'll send you the doc.
On Thu, Apr 14, 2016 at 5:02 PM, Jay Potharaju <jspothar...@gmail.com>
wrote:
> Thanks John, which version of solr are you using?
>
> On Thu, Apr 14, 2016 at 3:59 PM, John Bickerstaff <
> j..
> Curious what command did you use?
>
> On Thu, Apr 14, 2016 at 3:48 PM, John Bickerstaff <
> j...@johnbickerstaff.com>
> wrote:
>
> > I had a hard time getting replicas made via the API, once I had created
> the
> > collection for the first time although t
I had a hard time getting replicas made via the API, once I had created the
collection for the first time although that may have been ignorance on
my part.
I was able to get it done fairly easily on the Linux command line. If
that's an option and you're interested, let me know - I have a
t
> boosting just influences the score it does _not_ explicitly order the
> results. So the docs with "figo" in the conentType field will tend to
> the top, but won't be absolutely guaranteed to be there.
>
>
>
> Best,
> Erick
>
> On Thu, Apr 14, 2016 at 12:18 PM
terms in solrconfig.xml. You've already found the ability
> to configure edismax as your defType and apply boosts
> to particular fields...
>
> Best,
> Erick
>
> On Thu, Apr 14, 2016 at 11:53 AM, John Bickerstaff
> <j...@johnbickerstaff.com> wrote:
> >
.
Thanks.
On Thu, Apr 14, 2016 at 12:34 PM, John Bickerstaff <j...@johnbickerstaff.com
> wrote:
> I have the following (essentially hard-coded) line in the Solr Admin Query
> UI
>
> =
> bq: contentType:(searchTerm1 searchTerm2 searchTerm2)^1000
> =
>
> The
I have the following (essentially hard-coded) line in the Solr Admin Query
UI
=
bq: contentType:(searchTerm1 searchTerm2 searchTerm2)^1000
=
The "searchTerm" entries represent whatever the user typed into the search
box. This can be one or more words. Usually less than 5.
I want to
Stupid phone autocorrect...
If you add updated documents of the same ID over time, optimizing your
collection(s) may help.
On Thu, Apr 14, 2016 at 7:50 AM, John Bickerstaff <j...@johnbickerstaff.com>
wrote:
> If you delete a lot of documents over time, or if you add updated
&g
If you delete a lot of documents over time, or if you add updated documents
of the same I'd over time, optimizing your collection(s) may help.
On Apr 14, 2016 3:52 AM, "Emir Arnautovic"
wrote:
> Hi Edwin,
> Indexing speed depends on multiple factors: HW, Solr
Hello all,
I'm wondering if anyone can comment on arguments for and against putting
solr.xml into Zookeeper?
I assume one argument for doing so is that I would then have all
configuration in one place.
I also assume that if it doesn't get included as part of the upconfig
command, there is
t; Thanks
>
> Bosco
>
>
>
>
> On 4/6/16, 2:47 PM, "John Bickerstaff" <j...@johnbickerstaff.com> wrote:
>
> >Therefore, this becomes possible:
> >
> http://stackoverflow.com/questions/525212/how-to-run-unix-shell-script-from-java-code
> >
&
Therefore, this becomes possible:
http://stackoverflow.com/questions/525212/how-to-run-unix-shell-script-from-java-code
Hackish, but certainly doable... Given there's no API...
On Wed, Apr 6, 2016 at 3:44 PM, John Bickerstaff <j...@johnbickerstaff.com>
wrote:
> Yup - just tested - tha
Yup - just tested - that command runs fine with Solr NOT running...
On Wed, Apr 6, 2016 at 3:41 PM, John Bickerstaff <j...@johnbickerstaff.com>
wrote:
> If you can get to the IP addresses from your application, then there's
> probably a way... Do you mean you're firewalled off or i
If you can get to the IP addresses from your application, then there's
probably a way... Do you mean you're firewalled off or in some other way
unable to access the Solr box IP's from your Java application?
If you're looking to do "automated build of virtual machines" there are
some tools like
Right... You can store that anywhere - but at least consider not storing
it in your existing SOLR collection just because it's there... It's not
really the same kind of data -- it's application meta-data and/or
user-specific data...
Getting it out later will be more difficult than if you store
ties as per 5.5
> changes
> 5. start zookeeper
> 5. upload config to zookeeper
> 6. Create collection using rest api
> 7. start cluster
> 8. copy collection data from 4.5 to solr 5.5 data directory
>
>
> If you can share upgrade step/process document, that will be great
>
/confluence/display/solr/Upgrading+Solr
https://cwiki.apache.org/confluence/display/solr/Upgrading+a+Solr+4.x+Cluster+to+Solr+5.0
https://cwiki.apache.org/confluence/display/solr/Major+Changes+from+Solr+4+to+Solr+5
On Wed, Apr 6, 2016 at 8:58 AM, John Bickerstaff <j...@johnbickerstaff.com>
wrote
I recently upgraded from 4.x to 5.5 -- it was a pain to figure it out, but
it turns out to be fairly straightforward...
Caveat: Because I run all my data into Kafka first, I was able to easily
re-create my collections by running a microservice that pulls from Kafka
and dumps into Solr.
I have a
A few thoughts...
>From a black-box testing perspective, you might try changing that
softCommit time frame to something longer and see if it makes a difference.
The size of your documents will make a difference too - so the comparison
to 300 - 500 on other cloud setups may or may not be
In terms of #2, this might be of use...
https://wiki.apache.org/solr/HowToReindex
On Tue, Apr 5, 2016 at 3:08 PM, Anuj Lal wrote:
> I am new to solr. Need some advice from more experienced solr team
> members
>
> I am upgrading 4.4 solr cluster to 5.5
>
>
> One of the
>From some docs I'm working on - this command (against one solr box) got me
the entire cluster's state...
Don't know if it'll work for you, but just in case... There may be an api
command that is similar - not sure. I'm mostly operating on the command
line right now.
(statdx is the name of my
My own choices were driven mostly by the usage of the data - from a more
architectural perspective.
I have "appDocuments" and "appImages" for one of the applications I'm
supporting. Because they are so closely connected (an appDocuments can
have N number of appImages and appImages can belong to
ID's are on documents that are actually unique.
On Mon, Apr 4, 2016 at 9:51 PM, John Bickerstaff <j...@johnbickerstaff.com>
wrote:
> Sweet - that's a good point - I ran into that too - I had not run the
> commit for the last "batch" (I was using SolrJ) and so numbers didn
Sweet - that's a good point - I ran into that too - I had not run the
commit for the last "batch" (I was using SolrJ) and so numbers didn't match
until I did.
On Mon, Apr 4, 2016 at 9:50 PM, Binoy Dalal wrote:
> 1) Are you sure you don't have duplicates?
> 2) All of your
The first question is whether you have duplicate ID's in your data set.
I had the same kind of thing a few months back, freaked out, and spent a
few hours trying to figure it out by coding extra logging etc... to keep
track of every single count at every stage of the process.. All the
numbers
Does SOLR cloud push indexing across all nodes? I've been planning 4 SOLR
boxes with only 3 exposed via the load balancer, leaving the 4th available
internally for my microservices to hit with indexing work.
I was assuming that if I hit my "solr4" IP address, only "solr4" will do
the indexing...
Will the processes be Solr processes? Or do you mean multiple threads
hitting the same Solr server(s)?
There will be a natural bottleneck at one Solr server if you are hitting it
with a lot of threads - since that one server will have to do all the
indexing.
I don't know if this idea is
You can sort like this (I believe that _version_ is the internal id/index
number for the document, but you might want to verify)
In the Admin UI, enter the following in the sort field:
_version_ asc
You could also put an entry in the default searchHandler in solrconfig.xml
to do this to every
I can display a sorted list
> via:
>
> fq=listid_s:378
> sort=listpos(listpos_s,378) asc
>
> Regards,
> Tamas
>
> On Fri, Apr 1, 2016 at 8:55 PM, John Bickerstaff <j...@johnbickerstaff.com
> >
> wrote:
>
> > Tamas,
> >
> > This feels a bit li
l...
http://stackoverflow.com/questions/3931827/solr-merging-results-of-2-cores-into-only-those-results-that-have-a-matching-fie
On Fri, Apr 1, 2016 at 12:40 PM, John Bickerstaff <j...@johnbickerstaff.com>
wrote:
> Tamas,
>
> I'm brainstorming here - not being careful, just throwing o
-
and then sort them by ID based on the data associated with the User (a list
of ID's, in order)
There is even a way to write a plugin that will go after external data to
help sort Solr documents, although I'm guessing you'd rather avoid that...
On Fri, Apr 1, 2016 at 11:59 AM,
(status, amount, ..) from offset and 50
> rows then it would be perfect and fast. If ordering would be outside of
> solr then i have to retrive almost every 1 documents from solr (a bit
> less if filtered) to order them and display the page of 50 products.
> 2016. ápr. 1. 19:15 ez
ver your intent is for this search.
On Fri, Apr 1, 2016 at 11:15 AM, John Bickerstaff <j...@johnbickerstaff.com>
wrote:
> Just to be clear - I don't mean who requests the list (application or
> user) I mean what "rule" determines the ordering of the list?
>
> Or, is there even a
ecause Solr does the work of
> filtering and pagination. If sorting were done outside than I would have to
> read every document from Solr to sort them. It is not an option, I have to
> query onle one page.
>
> I don't understand how to solve it using subqueries.
> 2016. ápr
Specifically, what drives the position in the list? Is it arbitrary or is
it driven by some piece of data?
If data-driven - code could do the sorting based on that data... separate
from SOLR...
Alternatively, if the data point exists in SOLR, a "sub-query" might be
used to get the right sort
I believe I want to set up a search handler with a function query to avoid
needing to code it.
The function query does some weighting by checking the "title" field for
whatever the user entered as their search term (named myCurrentSearchTerm
below)
To test this out in the Admin UI, I have the
201 - 248 of 248 matches
Mail list logo