If you want to override facet.field through the query, you have to override ALL
facet.field's defined as default in reqeust handler, else those other facets
are gone.
You say But it's not working. without specifying WHAT is not working.
--
Jan Høydahl, search solution architect
Cominvent AS -
Hi,
Exists any way to eject a node from a solr cluster?
If I shutdown a node in the cluster, the zookeeper tag the node as down.
Thanks
/Yago
-
Best regards
--
View this message in context:
http://lucene.472066.n3.nabble.com/Eject-a-node-from-SolrCloud-tp4038950.html
Sent from the
Yes, currently the only option is to shutdown the node. Maybe not the
cleanest way to remove a node. See this jira too:
https://issues.apache.org/jira/browse/SOLR-3512
On Thu, Feb 7, 2013 at 7:20 AM, yriveiro yago.rive...@gmail.com wrote:
Hi,
Exists any way to eject a node from a solr
Hi Soumyanayan,
We developed a parser that converts dtSearch queries to Lucene queries, with
some Solr integration - see
http://www.flax.co.uk/blog/2012/04/24/dtsolr-an-open-source-replacement-for-the-dtsearch-closed-source-search-engine/
At the moment it relies on an unreleased version of
How was your facet.field defined in the request handler? My guess is it needs
to be moved to an appends section.
Erik
On Feb 7, 2013, at 4:11, Jan Høydahl jan@cominvent.com wrote:
If you want to override facet.field through the query, you have to override
ALL facet.field's
Hi
I need configure a mysql pool connection on Solr Server for using on
custom plugin. I saw DataImportHandler wiki:
http://wiki.apache.org/solr/DataImportHandler , but it's seems that
DataImportHandler open the connection when handler is calling and close
when finish import and I need
Rafal,
What about docnums, don't they are limited by int32 ?
07.02.2013 15:33 пользователь Rafał Kuć r@solr.pl написал:
Hello!
Practically there is no limit in how many documents can be stored in a
single index. In your case, as you are using Solr from 2011, there is
a limitation
Changing x to y (e.g., s2 to sII) is not a function of spell check
or suggestion.
Synonyms are a closer match, but can be difficult to configure properly.
Good luck.
You may be better off preprocessing the query at the application level and
then generating the appropriate boolean logic,
Hello!
Right, my bad - ids are still using int32. However, that still
gives us 2,147,483,648 possible identifiers right per single index,
which is not close to the 13,5 millions mentioned in the first mail.
--
Regards,
Rafał Kuć
Sematext :: http://sematext.com/ :: Solr - Lucene - Nutch -
You can unload the core for that node and it will be removed from zookeeper.
You can add it back after if you leave it's state on disk and recreate the core.
- Mark
On Feb 7, 2013, at 5:20 AM, yriveiro yago.rive...@gmail.com wrote:
Hi,
Exists any way to eject a node from a solr cluster?
Hi,
My schema is like below
fields
field name=ProductId type=int indexed=true stored=true /
field name=ProductName type=string indexed=true stored=true
required=true /
field name=ProductDesription type=string indexed=true stored=true
required=true /
field
Hi,
My schema is like below
fields
field name=ProductId type=int indexed=true stored=true /
field name=ProductName type=text_general indexed=true stored=true
required=true /
field name=ProductDesription type=string indexed=true stored=true
required=true /
field
I think it would still fail because of the 'tag' exclusions. Whatever
facets will be defined on the handler, they will not be taking into account
'tag' exclusions.
Regards,
Alex.
Personal blog: http://blog.outerthoughts.com/
LinkedIn: http://www.linkedin.com/in/alexandrerafalovitch
- Time is
Thanks, Mark. I created SOLR-4413 [1] for it.
I'm not sure what the best fix is since it looks like a lot of the work at
that time went into refactoring SolrIndexSearcher to use DirectoryFactory
everywhere and index.properties doesn't make much sense when an FSDirectory
is not used...
Anyway,
Hi,
I am no Korean expert and am finding it difficult to fix this, my client is
searching for the following query, but the NOT condition doesn't seem to be
working.
(stnostem:((옵티머스 OR 엘지 스마트폰) AND NOT (옵티머스 프라임 OR 프라임)))
the search result (xml attached ) return result with the not condition
I'm looking into a way to implement an autosuggest and for my special needs
(I'm doing a startsWith-search that should retrieve the full name, which
may have accents - However, I want to search with/without accents and in
any upper/lowercase for comfort)
Here's part of my configuration:
Actually, I have a dream to exceed those two billions. It seems possible,
to move to Vint in fileformat and change int docnums to longs in Lucene
API. Does anyone know whether it's possible?
And this question is not so esoteric if we are talking about SolrCloud,
which can hold more that 2bn docs
I'm not very familiar with how AnalyzingSuggester works inside Solr
... if you try this directly with the Lucene APIs does it still
happen?
Hmm maybe one idea: if you remove whitespace from your suggestion does
it work? I wonder if there's a whitespace / multi-token issue ... if
so then maybe
Hi
I need configure a mysql pool connection on Solr Server for using on
custom plugin. I saw DataImportHandler wiki:
http://wiki.apache.org/solr/DataImportHandler , but it's seems that
DataImportHandler open the connection when handler is calling and close
when finish import and I need
On 2013-02-07 14:58, Pragyanshis Pattanaik wrote:
Hi,
Hi,
My schema is like below
fields
field name=ProductId type=int indexed=true stored=true /
field name=ProductName type=string indexed=true stored=true
required=true /
field name=ProductDesription type=string indexed=true
It's the same with whitespace removed unfortunately - still getting back
testname then.
I'm not quite sure how to test this via the Lucene API - in particular, how
to define the KeywordTokenizer with ASCII+LowerCase, so I can't test this
atm :/
BR Sebastian Saip
On 7 February 2013 16:19,
Hello Miguel,
If you set up a JNDI datasource in your servlet container, you can use
that as your database config. Then you just need to use a pooling
datasource:
http://wiki.apache.org/solr/DataImportHandlerFaq#How_do_I_use_a_JNDI_DataSource.3F
Small addition:
To support query, I probably have to implement an analyzer (query time)...
An analyzer can be configured on numeric (i.e non TEXT) field?
On Thu, Feb 7, 2013 at 6:48 PM, Isaac Hebsh isaac.he...@gmail.com wrote:
Hi.
I have to index field which contains an IP address.
Users
http://wiki.apache.org/solr/ExtendedDisMax
On Thu, Feb 7, 2013 at 6:53 PM, Pragyanshis Pattanaik
pragyans...@outlook.com wrote:
Thanks for the reply
But i need to boost each field so the first approach might not be
applicable here right ?
Date: Thu, 7 Feb 2013 06:08:25 -0800
From:
Hello Alex,
I asked a similar question on server fault:
http://serverfault.com/a/474442/156440
On Wed, Feb 6, 2013 at 7:05 PM, Alexandre Rafalovitch arafa...@gmail.comwrote:
Hello,
When I CTRL-C the example Solr, it prints a bunch of graceful shutdown
messages. I assume it shuts down safe
The solution, as pointed out on
http://stackoverflow.com/questions/14732713/solr-autosuggest-with-diacritics/14743278
,
is not to use a copyField but instead use the AnalyzingSuggester on the
StrField directly.
Cheers!
On 7 February 2013 17:30, Sebastian Saip sebastian.s...@gmail.com wrote:
Glad it helped. :-)
Now, if you could write this up as a full example and explanation, I am
sure Solr community would benefit from it as well. If you don't have your
own blog, I would be happy to guest host it, as I am sure would at least a
couple more people/organizations.
Regards,
Alex.
Good day,
I got my elevation component working with the /elevate handler.
However, I would like to add the elevation component to my main search
handler which is currently /query.
so I can have one handler return everything (elevated items with regular
search results; i.e. one stop shopping, so
Thanks, but it is not quite the same. I am talking about SolrJ, where Solr
is hosted within an application, not in a servlet container.
Regards,
Alex.
Personal blog: http://blog.outerthoughts.com/
LinkedIn: http://www.linkedin.com/in/alexandrerafalovitch
- Time is the quality of nature that
Hi,
I think yes CoreContainer.shutdown() is the right method for embedded solr
server.
But i believe embedded solr server is not preferred anymore after javabin
codec. And it is not well tested compared to http server.
--- On Thu, 2/7/13, Alexandre Rafalovitch arafa...@gmail.com wrote:
Update:
Ok, If I search for gangnam style in /query handler by itself, elevation
works!
If I search with gangnam style and/or something else the elevation component
doesn't work but the rest of the query does.
here's the examples:
works:
Howdy,
I have a Solr implementation that allows me to do a geospatial search and
I'm trying to replicate it using the solrj libs. The schema.xml that I'm
using looks like this:
schema name=v3_geo version=1.1
types
fieldtype name=string class=solr.StrField sortMissingLast=true
For some reason I thought Solr 4.0 brought us the ability to sort groups by
numFound when using field collapsing, but now I can't find any info about how
to do that. Did I make it all up? Might I be able to fake it using pseudo
fields somehow?
Thanks,
Tony
Hi Solr community
I'm in the process of getting my mind set straight on SolrCloud; more
specifically: trying to design a feasible workflow for a use-case where we
currently use master/slave replication. First, the use case:
We want to
1. separate indexing workload from query workload
2.
Looks like SolrServer.shutdown() is the more standard approach. For
embedded server, it just calls container.shutdown() anyway.
Regards,
Alex.
Personal blog: http://blog.outerthoughts.com/
LinkedIn: http://www.linkedin.com/in/alexandrerafalovitch
- Time is the quality of nature that keeps
On Feb 7, 2013, at 5:23 PM, Steffen Elberg Godskesen
steffen.godske...@gmail.com wrote:
Re (1): Am I right in thinking that a given update is sent to every replica
of the shard to which it belongs for analysis and indexing? And that there is
no immediate way to separate indexing from
You should run replicated ZK:
http://zookeeper.apache.org/doc/trunk/zookeeperStarted.html#sc_RunningReplicatedZooKeeper
Give Solr the list of all ZK's and you're good to go
--
Jan Høydahl, search solution architect
Cominvent AS - www.cominvent.com
Solr Training - www.solrtraining.com
7. feb.
Hi all,
I`m struggiling with strange local params behaviour when {!tag=somethig}
is in brackets term.
case:
facet.field={!ex=d0feea8}categoryfq={!tag=d0feea8}category:5 AND
type:DOCUMENT
works as I expected, but:
facet.field={!ex=d0feea8}categoryfq=({!tag=d0feea8}category:5 OR
I'm pretty sure the local params have to be at the very start of the query. But
you should be able to do this with nested queries. Try this...
fq=_query_:{!tag=d0feea8}category:\5\ OR otherField:\otherValue\ AND
type:DOCUMENT
-Michael
-Original Message-
From: Karol Sikora
Hi all,
Is there any parameter which will set the number of document returned
after applying grouping on results? Like we have query.setRows for results
without grouping?
I know all the below functions will set to group param. But this will not
limit number of document returned. We want to
this was my previous data.[
{
id:1,
movie_name:Twelve Monkeys,
genre:[
Children's,
Comedy,
Drama,
Sci-Fi
],
release_year:1995,
url:http://us.imdb.com/M/title-exact?Twelve%20Monkeys%20(1995),
rating:8.1,
Hi Rajani,
I recently tried to solve a similar problem as the one you have. (I think)
Solr doesn't support a param to achieve this because if we were to limit
the no of documents returned, to get the next result set the starting
offset of each group will be different based on the number of
It is giving zero response mostly because your default query operator is
set to AND.
try this:
fq={!lucene q.op=OR}institute_name:(xyz sfsda sdfsaf)
On 8 February 2013 12:26, anurag.jain anurag.k...@gmail.com wrote:
Not working through that style. it is giving me zero response
--
View
On 2/7/2013 9:29 PM, Alexandre Rafalovitch wrote:
Hello,
What actually happens when using soft (as opposed to hard) commit?
I understand somewhat very high-level picture (documents become available
faster, but you may loose them on power loss).
I don't care about low-level implementation
44 matches
Mail list logo