Re: Error occurred when getting solr-core

2014-11-24 Thread Danesh Kuruppu
Hi all,

Cores are not loaded to the CoreContainer. my solr.xml configuration is

solr
   solrcloud
 str name=host127.0.0.1/str
 int name=hostPort${jetty.port:8983}/int
 str name=hostContext${hostContext:solr}/str
 int name=zkClientTimeout${zkClientTimeout:3}/int
 str name=shareSchema${shareSchema:false}/str
 bool name=genericCoreNodeNames${genericCoreNodeNames:true}/bool
   /solrcloud

   shardHandlerFactory name=shardHandlerFactory
 class=HttpShardHandlerFactory
 int name=socketTimeout${socketTimeout:12}/int
 int name=connTimeout${connTimeout:15000}/int
   /shardHandlerFactory
 /solr


Please help.
Thanks

On Mon, Nov 24, 2014 at 11:35 AM, Danesh Kuruppu dknkuru...@gmail.com
wrote:

 Hi all,

 I am using solr version 4.7.2

 I am getting following when communicating with solr server.

 org.apache.solr.common.SolrException: No such core: db
 at
 org.apache.solr.client.solrj.embedded.EmbeddedSolrServer.request(EmbeddedSolrServer.java:112)
 at
 org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:118)
 at org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:116)
 at org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:102)


 I am using EmbeddedSorlServer, It is initialized as follows

 CoreContainer coreContainer = new
 CoreContainer(solrHome.getPath());
 coreContainer.load();
 this.server = new EmbeddedSolrServer(coreContainer, db);


 in solr-home, there is a directory called db. in core.properties file, I
 define the name and dataDir.

 Couldn't find the cause for this error.

 Please help.

 Thanks
 Danesh



Reminder: FOSDEM 2015 - Open Source Search Dev Room

2014-11-24 Thread Uwe Schindler
Hi,

We host a Dev-Room about Open Source Search on this year's FOSDEM 2015 
(https://fosdem.org/2015/), taking place on January 31th and February 1st, 
2015, in Brussels, Belgium. There is still one more week to submit your talks, 
so hurry up and submit your talk early!

Here is the full CFP as posted a few weeks ago:

Search has evolved to be much more than simply full-text search. We now rely on 
“search engines” for a wide variety of functionality:
search as navigation, search as analytics and backend for data visualization 
and sometimes, dare we say it, as a data store. The purpose of this dev room is 
to explore the new world of open source search engines: their enhanced 
functionality, new use cases, feature and architectural deep dives, and the 
position of search in relation to the wider set of software tools.

We welcome proposals from folks working with or on open source search engines 
(e.g. Apache Lucene, Apache Solr, Elasticsearch, Seeks, Sphinx, etc.) or 
technologies that heavily depend upon search (e.g.
NoSQL databases, Nutch, Apache Hadoop). We are particularly interested in 
presentations on search algorithms, machine learning, real-world 
implementation/deployment stories and explorations of the future of search.

Talks should be 30-60 minutes in length, including time for QA.

You can submit your talks to us here:
https://docs.google.com/forms/d/11yLMj9ZlRD1EMU3Knp5y6eO3H5BRK7V38G0OxSfp84A/viewform

Our Call for Papers will close at 23:59 CEST on Monday, December 1, 2014. We 
cannot guarantee we will have the opportunity to review submissions made after 
the deadline, so please submit early (and often)!

Should you have any questions, you can contact the Dev Room
organizers: opensourcesearch-devr...@lists.fosdem.org

Cheers,
LH on behalf of the Open Source Search Dev Room Program Committee*

* Boaz Leskes, Isabel Drost-Fromm, Leslie Hawthorn, Ted Dunning, Torsten Curdt, 
Uwe Schindler

-
Uwe Schindler
uschind...@apache.org 
Apache Lucene PMC Member / Committer
Bremen, Germany
http://lucene.apache.org/




Re: Error occurred when getting solr-core

2014-11-24 Thread Danesh Kuruppu
Hi all,

Found the RC for this error. I missed some dependency library for sore-core.

Thanks
Danesh

On Mon, Nov 24, 2014 at 1:43 PM, Danesh Kuruppu dknkuru...@gmail.com
wrote:

 Hi all,

 Cores are not loaded to the CoreContainer. my solr.xml configuration is

 solr
   solrcloud
 str name=host127.0.0.1/str
 int name=hostPort${jetty.port:8983}/int
 str name=hostContext${hostContext:solr}/str
 int name=zkClientTimeout${zkClientTimeout:3}/int
 str name=shareSchema${shareSchema:false}/str
 bool name=genericCoreNodeNames${genericCoreNodeNames:true}/bool
   /solrcloud

   shardHandlerFactory name=shardHandlerFactory
 class=HttpShardHandlerFactory
 int name=socketTimeout${socketTimeout:12}/int
 int name=connTimeout${connTimeout:15000}/int
   /shardHandlerFactory
 /solr


 Please help.
 Thanks

 On Mon, Nov 24, 2014 at 11:35 AM, Danesh Kuruppu dknkuru...@gmail.com
 wrote:

 Hi all,

 I am using solr version 4.7.2

 I am getting following when communicating with solr server.

 org.apache.solr.common.SolrException: No such core: db
 at
 org.apache.solr.client.solrj.embedded.EmbeddedSolrServer.request(EmbeddedSolrServer.java:112)
 at
 org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:118)
 at org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:116)
 at org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:102)


 I am using EmbeddedSorlServer, It is initialized as follows

 CoreContainer coreContainer = new
 CoreContainer(solrHome.getPath());
 coreContainer.load();
 this.server = new EmbeddedSolrServer(coreContainer, db);


 in solr-home, there is a directory called db. in core.properties file, I
 define the name and dataDir.

 Couldn't find the cause for this error.

 Please help.

 Thanks
 Danesh





Re: Too much data after closed for HttpChannelOverHttp

2014-11-24 Thread Hakim Benoudjit
It seems to be working now.

- I have increased some values in jetty.xml:


*Set name=responseBufferSize65536/Set Set
name=headerBufferSize32768/Set *
*Set name=responseBufferSize32768/Set *

- And have left chunk size = 3000.
- But I'm not adding all the documents at the same time; instead items are
sent one by one to be indexed (it's either that or the jetty configuration
altered which is preventing the previous error from being reproduced).

Thanks everyone.

2014-11-24 2:36 GMT+01:00 Alexandre Rafalovitch arafa...@gmail.com:

 Good point on that one Steve.

 Wireshark is both a hammer and a power drill of network
 troubleshooting. Takes steady hands to hold it right (it has a bit of
 a learning curve) but it is a great tool. I swore by it (well Ethereal
 back then) in my tech support days.

 So, seconded to try using that if the simple approach fails outright.

 Regards,
Alex.
 Personal: http://www.outerthoughts.com/ and @arafalov
 Solr resources and newsletter: http://www.solr-start.com/ and @solrstart
 Solr popularizers community: https://www.linkedin.com/groups?gid=6713853


 On 23 November 2014 at 20:31, steve sc_shep...@hotmail.com wrote:
 
 
 
  For what it's worth, depending on the type of PC/MAC you're using, you
 can use WireShark to look at active http header (sent and received) that
 are being created for the request.
  https://www.wireshark.org/
  I don't have any financial interest in them, but the stuff works!
  Steve
 
  Date: Sun, 23 Nov 2014 20:47:05 +0100
  Subject: Re: Too much data after closed for HttpChannelOverHttp
  From: h.benoud...@gmail.com
  To: solr-user@lucene.apache.org
 
  Actually I'm using a php client (I think it sends a HTTP request to
 Solr),
  but you're right tomorrow once I'll get to the office, I'll set chunk
 size
  to a smaller value, and will tell you if that was the reason.
 
  Thanks.
 
  2014-11-23 19:35 GMT+01:00 Alexandre Rafalovitch arafa...@gmail.com:
 
   Most probably just a request that's too large. Have you tried dropping
   down to 500 items and seeing what happens?
  
   Are you using SolrJ to send content to Solr? Or a direct HTTP request?
  
   Regards,
  Alex.
   P.s. You may also find it useful to read up on the Solr commit and
   hard vs. soft commits. Check solrconfig.xml in the example
   distribution.
   Personal: http://www.outerthoughts.com/ and @arafalov
   Solr resources and newsletter: http://www.solr-start.com/ and
 @solrstart
   Solr popularizers community:
 https://www.linkedin.com/groups?gid=6713853
  
  
   On 23 November 2014 at 12:31, Hakim Benoudjit h.benoud...@gmail.com
   wrote:
Hi there,
   
I have deployed solr with Jetty, and I'm trying to index a quite
 large
amount of items (300K), retreived from a MySQL database
 (unfortunately
   I'm
not using DIH; I'm doing it manually, by getting items from MySQL
 and
   then
index them it in Solr).
   
But, I'm not indexing all of those items at the same time; I'm
 indexing
them by chunks of 3K.
So, I get the first 3K, index them, then goes to the next 3K chunk
 to
   index
it.
   
Here is the error I got in jetty logs, I guess it has nothing to do
 with
Mysql:
*Does anyone know the meaning of the error 'badMessage:
java.lang.IllegalStateException: too much data after closed for
HttpChannelOverHttp@5432494a' ?*
   
Thanks for your help, if anything isnt very precise please tell me
 to
explain it (and sorry for my bad english).
   
--
Cordialement,
Best regards,
Hakim Benoudjit
  
 
 
 
  --
  Cordialement,
  Best regards,
  Hakim Benoudjit
 
 




-- 
Cordialement,
Best regards,
Hakim Benoudjit


Re: Duplicate facets when the handler configuration specifies facet fields

2014-11-24 Thread Burke, Brian
Good to hear you can reproduce it and thank you for entering that issue into 
JIRA.  FWIW, here is the full list of parameters from echoParams=all from my 
test:


params:{
  mlt.minwl:2,
  mm:5,
  facet:true,
  mlt.boost:true,
  mlt.fl:title,summary,body,
  facet.mincount:1,
  mlt.mintf:2,
  f.typedef.facet.limit:15,
  mlt.qf:title^5 summary^3 body^1,
  mlt.maxqt:8,
  f.subtype.facet.limit:15,
  mlt.mindf:3,
  echoParams:all,
  df:id,
  mlt.count:3,
  facet.field:primaryId,
  facet:true,
  echoParams:all,
  indent:true,
  q:*:*,
  facet.limit:10,
  facet.field:primaryId,
  wt:json,
  rows:0}},

Also, our testing is for an upgrade from solr 4.4 to solr 4.10.2.  I found that 
in our current production 4.4 instance, the params shows the duplicate 
(facet.field listed twice), but the actual results only has one set of facets.  
Guessing there must have been some code on the output side that was removing 
the duplicates in that version of solr.

Thanks,
Brian


On Nov 24, 2014, at 6:04 AM, Alexandre Rafalovitch 
arafa...@gmail.commailto:arafa...@gmail.com wrote:

I can reproduce it.

I added your parameters to the default section of the config and then
run the following:
curl 
http://localhost:8983/solr/schemaless/select?q=*:*rows=0wt=jsonindent=truefacet=truefacet.field=primaryId2facet.limit=10echoParams=all;

I get:
--
  params:{
 f.typedef.facet.limit:15,
 facet.field:primaryId2,
 df:_text,
 f.subtype.facet.limit:15,
 echoParams:all,
 facet.mincount:1,
 rows:0,
 facet:true,
 q:*:*,
 facet.limit:10,
 facet.field:primaryId2,
 indent:true,
 echoParams:all,
 rows:0,
 wt:json,
 facet:true}},
--

This is against Solr 5 build, but I think that bug is there all the
way to Solr 4.1. I think I traced the source of the bug too (parameter
and default names are just joined together but with the first
(override) value both times as shown above). Usually makes no
difference to anything, but it looks like faceting component iterates
over the elements, not just gets them, so it gets bitten twice.

I've created a JIRA for this issue:
https://issues.apache.org/jira/browse/SOLR-6780

Regards,
  Alex.
Personal: http://www.outerthoughts.com/ and @arafalov
Solr resources and newsletter: http://www.solr-start.com/ and @solrstart
Solr popularizers community: https://www.linkedin.com/groups?gid=6713853


On 21 November 2014 at 18:29, Alexandre Rafalovitch arafa...@gmail.com wrote:
Could you add echoParams=all to the query and see what comes back?
Currently, you echo the params you sent, would be good to see what
they look like after they combine with defaults.

Regards,
 Alex.
Personal: http://www.outerthoughts.com/ and @arafalov
Solr resources and newsletter: http://www.solr-start.com/ and @solrstart
Solr popularizers community: https://www.linkedin.com/groups?gid=6713853


On 21 November 2014 18:04, Tom Zimmermann zimm.to...@gmail.com wrote:
Brian and I are working together to diagnose this issue so I can chime in
quickly here as well. These values are defined as part of the the defaults
section of the config.



Highlight returning multiValued field as a single string

2014-11-24 Thread xan
I've a multiValued field. Querying on it returns a multivalued result set.
But, doing a highlight query returns a single string. I also have
hl.preserveMulti=true option.

the schema of the field is as follows:

field name=some_field type=some_field_type indexed=true stored=true
multiValued=true termVectors=true termPositions=true
termOffsets=true /

fieldType name=some_field_type class=solr.TextField
positionIncrementGap=100
analyzer type=index
tokenizer class=solr.WhitespaceTokenizerFactory/

filter class=solr.WordDelimiterFilterFactory
generateWordParts=1 generateNumberParts=1 catenateWords=1
catenateNumbers=1 catenateAll=0 splitOnCaseChange=1 /
filter class=solr.LowerCaseFilterFactory/
/analyzer

analyzer type=query
tokenizer class=solr.WhitespaceTokenizerFactory/
filter class=solr.SynonymFilterFactory
synonyms=synonyms.txt ignoreCase=true expand=true/
filter class=solr.StopFilterFactory  
ignoreCase=true
words=lang/stopwords_en.txt
enablePositionIncrements=true
/
filter class=solr.WordDelimiterFilterFactory
generateWordParts=1 generateNumberParts=1 catenateWords=0
catenateNumbers=0 catenateAll=0 splitOnCaseChange=1/
filter class=solr.LowerCaseFilterFactory/
/analyzer
/fieldType



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Highlight-returning-multiValued-field-as-a-single-string-tp4170587.html
Sent from the Solr - User mailing list archive at Nabble.com.


RE: Lucene ancient greek normalization

2014-11-24 Thread Allison, Timothy B.
If you are using Solr, you can configure your analysis chain to use the 
ICUFoldingFilterFactory 
(https://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters#solr.ICUFoldingFilterFactory)
 and then view the results in the solr admin window.

If you are in pure Lucene (circa version 4.8, some mods will be required 
depending on your version):
1) Extend Analyzer:
@Override
protected TokenStreamComponents createComponents(String field, Reader 
reader) {
Tokenizer stream = new StandardTokenizer(version, reader);
TokenFilter icu = new ICUFoldingFilter(stream);
return new TokenStreamComponents(stream, icu);
}

2)
Then iterate through the tokens:

TokenStream stream = analyzer.tokenStream(, new 
StringReader(text));
stream.reset();
CharTermAttribute cattr = 
stream.getAttribute(CharTermAttribute.class);
while (stream.incrementToken()) {
String token = cattr.toString();
...
-Original Message-
From: paolo anghileri [mailto:paolo.anghil...@codegeneration.it] 
Sent: Saturday, November 22, 2014 11:41 AM
To: Allison, Timothy B.
Subject: Re: Lucene ancient greek normalization

Sorry Timothy for the beginner question, how did you manage to run this 
test?

Many thanks

Paolo

On 21/11/2014 21:14, Allison, Timothy B. wrote:
 ICU looks promising:

 Μῆνιν ἄειδε, θεὰ, Πηληϊάδεω Ἀχιλλῆος -

 1.μηνιν
 2.αειδε
 3.θεα
 4.πηληιαδεω
 5.αχιλληοσ

 -Original Message-
 From: Alexandre Rafalovitch [mailto:arafa...@gmail.com]
 Sent: Friday, November 21, 2014 3:08 PM
 To: d...@lucene.apache.org
 Subject: Re: Lucene ancient greek normalization

 Are you sure that's not something that's already addressed by the ICU
 Filter? 
 http://www.solr-start.com/javadoc/solr-lucene/org/apache/lucene/analysis/icu/ICUTransformFilterFactory.html

 If you follow the links to what's possible, the page talks about
 Greek, though not ancient:
 http://userguide.icu-project.org/transforms/general#TOC-Greek

 There was also some discussion on:
 https://issues.apache.org/jira/browse/LUCENE-1343

 Regards,
 Alex.
 Personal: http://www.outerthoughts.com/ and @arafalov
 Solr resources and newsletter: http://www.solr-start.com/ and @solrstart
 Solr popularizers community: https://www.linkedin.com/groups?gid=6713853


 On 21 November 2014 14:14, paolo anghileri
 paolo.anghil...@codegeneration.it wrote:
 For development purposes I need the ability in lucene to normalize ancient
 greek characters for al the cases of grammatical details such as accents,
 diacritics and so on.

 My need is to retrieve ancient greek words with accents and other
 grammatical details by the input of the string without accents.

 For example the input of οργανον (organon) should to retrieve also  Ὄργανον,


 I am not a lucene commiter and I a new to this so my question is about the
 best practice to implement this in Lucene, and possibile submit a commit
 proposal to Lucene A project management committee.

 I have made some searches and found this file in Lucene-soir:


 It contains normalization for some chars.
 My thought would be to add extra normalization here, including all unicode
 ancient greek chars with all grammatical details.
 I already have all the unicode values for that chars so It should not be
 difficult for me to include them

 If my understanding is correct, this should add to lucene the features
 described above.


 As I am new to this, my needs are:

   To be sure that this is the correct place in Lucene for doing normalization
 How to post commit proposal


 Any help appreciated

 Kind regards

 Paolo
 -
 To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
 For additional commands, e-mail: dev-h...@lucene.apache.org


 -
 To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
 For additional commands, e-mail: dev-h...@lucene.apache.org




Re: Indexing problems with BBoxField

2014-11-24 Thread remus
OK, David Smiley now already created an issue for this:

https://issues.apache.org/jira/browse/SOLR-6781

So it really is a bug.
Furthermore, I also had a lot of problems trying to search on the field
after finally getting it indexed. I summarized those here:

https://issues.apache.org/jira/browse/SOLR-6784

On 2014-11-23 14:34, re...@gmx.net wrote:
 Thanks a lot for your reply!
 
 I had »docValues=true« in there before, but then thought I'd try out
 removing it to see if that helped. It didn't, and I forgot to re-add it
 before copying it into the mail.
 So, unfortunately, that's not it.
 
 However, the other one seems to bring us a step closer to the solution:
 After adding
 
 field name=bboxs_field_location_area type=bbox indexed=true
 stored=false multiValued=false/
 
 (even without removing the dynamic fields), this works indeed just fine!
 So, the question is what causes this, and it seems more and more like a
 bug instead of a user error. But I'll wait for a bit more feedback
 before filing a Jira.
 
 On 2014-11-23 14:10, Jack Krupansky wrote:
 A difference I see in your snippet from the example is that you don't
 have docValues=true on the coordinate field type. You wrote:

 fieldType name=_bbox_coord class=solr.TrieDoubleField
 precisionStep=8 stored=false /

 But the example is:

 fieldType name=_bbox_coord class=solr.TrieDoubleField
 precisionStep=8 docValues=true stored=false/

 Also, maybe try a static field rather than dynamic field, although the
 latter should work anyway.

 Please file a Jira to request that Solr give a user-sensible error, not
 a Lucene-level error. I mean, the Solr user has no ability to directly
 invoke the createFields method.

 And now... let's see what David Smiley has to say about all of this!

 -- Jack Krupansky

 -Original Message- From: Thomas Seidl
 Sent: Sunday, November 23, 2014 6:33 AM
 To: solr-user@lucene.apache.org
 Subject: Indexing problems with BBoxField

 Hi all,

 I just downloaded Solr 4.10.2 and wanted to try out the new BBoxField
 type, but couldn't get it to work. The error (with status 400) I get is:

 ERROR: [doc=foo] Error adding field
 'bboxs_field_location_area'='ENVELOPE(25.89, 41.13, 47.07, 35.31)'
 msg=java.lang.IllegalStateException: instead call createFields() because
 isPolyField() is true

 Which, of course, is rather unhelpful for a user.
 The relevant portions of my schema.xml look like this (largely copied
 from [1]:

 fieldType name=bbox class=solr.BBoxField geo=true units=degrees
 numberType=_bbox_coord /
 fieldType name=_bbox_coord class=solr.TrieDoubleField
 precisionStep=8 stored=false /
 dynamicField name=bboxs_* type=bbox indexed=true stored=false
 multiValued=false/

 [1] https://cwiki.apache.org/confluence/display/solr/Spatial+Search

 And the request I send is this:

 add
  doc
field name=idfoo/field
field name=bboxs_field_location_areaENVELOPE(25.89, 41.13,
 47.07, 35.31)/field
  /doc
 /add

 Does anyone have any idea what could be going wrong here?

 Thanks a lot in advance,
 Thomas

 


Re: Indexing problems with BBoxField

2014-11-24 Thread david.w.smi...@gmail.com
Thomas,

Thank you for communicating on the list about your experience and raising
the JIRA issue.  I meant to respond last night but lost the chance.  (and
Jack, thanks for helping Thomas out).  I’ll follow-up to SOLR-6784.
SOLR-6781 now has a bug-fix patch. I’ll apply it later today.

~ David Smiley
Freelance Apache Lucene/Solr Search Consultant/Developer
http://www.linkedin.com/in/davidwsmiley

On Mon, Nov 24, 2014 at 8:43 AM, re...@gmx.net wrote:

 OK, David Smiley now already created an issue for this:

 https://issues.apache.org/jira/browse/SOLR-6781

 So it really is a bug.
 Furthermore, I also had a lot of problems trying to search on the field
 after finally getting it indexed. I summarized those here:

 https://issues.apache.org/jira/browse/SOLR-6784

 On 2014-11-23 14:34, re...@gmx.net wrote:
  Thanks a lot for your reply!
 
  I had »docValues=true« in there before, but then thought I'd try out
  removing it to see if that helped. It didn't, and I forgot to re-add it
  before copying it into the mail.
  So, unfortunately, that's not it.
 
  However, the other one seems to bring us a step closer to the solution:
  After adding
 
  field name=bboxs_field_location_area type=bbox indexed=true
  stored=false multiValued=false/
 
  (even without removing the dynamic fields), this works indeed just fine!
  So, the question is what causes this, and it seems more and more like a
  bug instead of a user error. But I'll wait for a bit more feedback
  before filing a Jira.
 
  On 2014-11-23 14:10, Jack Krupansky wrote:
  A difference I see in your snippet from the example is that you don't
  have docValues=true on the coordinate field type. You wrote:
 
  fieldType name=_bbox_coord class=solr.TrieDoubleField
  precisionStep=8 stored=false /
 
  But the example is:
 
  fieldType name=_bbox_coord class=solr.TrieDoubleField
  precisionStep=8 docValues=true stored=false/
 
  Also, maybe try a static field rather than dynamic field, although the
  latter should work anyway.
 
  Please file a Jira to request that Solr give a user-sensible error, not
  a Lucene-level error. I mean, the Solr user has no ability to directly
  invoke the createFields method.
 
  And now... let's see what David Smiley has to say about all of this!
 
  -- Jack Krupansky
 
  -Original Message- From: Thomas Seidl
  Sent: Sunday, November 23, 2014 6:33 AM
  To: solr-user@lucene.apache.org
  Subject: Indexing problems with BBoxField
 
  Hi all,
 
  I just downloaded Solr 4.10.2 and wanted to try out the new BBoxField
  type, but couldn't get it to work. The error (with status 400) I get is:
 
  ERROR: [doc=foo] Error adding field
  'bboxs_field_location_area'='ENVELOPE(25.89, 41.13, 47.07, 35.31)'
  msg=java.lang.IllegalStateException: instead call createFields() because
  isPolyField() is true
 
  Which, of course, is rather unhelpful for a user.
  The relevant portions of my schema.xml look like this (largely copied
  from [1]:
 
  fieldType name=bbox class=solr.BBoxField geo=true units=degrees
  numberType=_bbox_coord /
  fieldType name=_bbox_coord class=solr.TrieDoubleField
  precisionStep=8 stored=false /
  dynamicField name=bboxs_* type=bbox indexed=true stored=false
  multiValued=false/
 
  [1] https://cwiki.apache.org/confluence/display/solr/Spatial+Search
 
  And the request I send is this:
 
  add
   doc
 field name=idfoo/field
 field name=bboxs_field_location_areaENVELOPE(25.89, 41.13,
  47.07, 35.31)/field
   /doc
  /add
 
  Does anyone have any idea what could be going wrong here?
 
  Thanks a lot in advance,
  Thomas
 
 



use {!surround} with more conditions

2014-11-24 Thread Bet Hel
Can't make surround work with other conditions...
for example:

{!surround}SomeField:30w(Text1*,Text2*)  -SomeField:Text3*

SomeField:* -{!surround}SomeField:30w(Text1*,Text2*)

Is that doable?


Missing value with Date Range

2014-11-24 Thread nabil Kouici

   
Hi All,
I'm trying to get missing count with Date Range adding facet.missing=true as 
parameter but this not work (which work with facet).
Do you have idea how to get missing count with date range or numeric range.
Regards,Nabil.

   

Backuping SolrCloud

2014-11-24 Thread elmerfudd
Hi, I'm looking for a built-in SolrCloud backup mechanism.
I want to backup my Index (scheduled / manual backups) while Indexing and
searching.

What is the proper way to perform this backup-restore task?


Thanks.
 



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Backuping-SolrCloud-tp4170624.html
Sent from the Solr - User mailing list archive at Nabble.com.


how solr4.0 and zookeeper run on weblogic

2014-11-24 Thread Sharma, Kunal
Hi ,

Is there a way to deploy Zookeeper on weblogic cluster.

Thanks and Regards
Kunal Sharma
07917836122



CoreContainer : create new cores reusing/sharing solrconfig.xml and schema.xml

2014-11-24 Thread Clemens Wyss DEV
Given the following directory/file structure

.../solrhome
.../solrhome/solr.xml
.../solrhome/config/
.../solrhome/config/schema.xml
.../solrhome/config/solrconfig.xml
.../solrhome/cores

How can I programmatically (CoreContainer) create a new core below 
.../solrhome/cores that makes use of solrconfig.xml and schema.xml? 
The new core should use the config files in .../solrhome/config/

Thx
Clemens


Re: CoreContainer : create new cores reusing/sharing solrconfig.xml and schema.xml

2014-11-24 Thread Alexandre Rafalovitch
Have you tried Configsets:
https://cwiki.apache.org/confluence/display/solr/Config+Sets ?

Regards,
   Alex.
Personal: http://www.outerthoughts.com/ and @arafalov
Solr resources and newsletter: http://www.solr-start.com/ and @solrstart
Solr popularizers community: https://www.linkedin.com/groups?gid=6713853


On 24 November 2014 at 10:14, Clemens Wyss DEV clemens...@mysign.ch wrote:
 Given the following directory/file structure

 .../solrhome
 .../solrhome/solr.xml
 .../solrhome/config/
 .../solrhome/config/schema.xml
 .../solrhome/config/solrconfig.xml
 .../solrhome/cores

 How can I programmatically (CoreContainer) create a new core below 
 .../solrhome/cores that makes use of solrconfig.xml and schema.xml?
 The new core should use the config files in .../solrhome/config/

 Thx
 Clemens


AW: CoreContainer : create new cores reusing/sharing solrconfig.xml and schema.xml

2014-11-24 Thread Clemens Wyss DEV
Thanks fort he quick reply. 
But how do I hand in the configSet to use. I guess through the CoreDescriptor?

-Ursprüngliche Nachricht-
Von: Alexandre Rafalovitch [mailto:arafa...@gmail.com] 
Gesendet: Montag, 24. November 2014 16:27
An: solr-user
Betreff: Re: CoreContainer : create new cores reusing/sharing solrconfig.xml 
and schema.xml

Have you tried Configsets:
https://cwiki.apache.org/confluence/display/solr/Config+Sets ?

Regards,
   Alex.
Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and 
newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers 
community: https://www.linkedin.com/groups?gid=6713853


On 24 November 2014 at 10:14, Clemens Wyss DEV clemens...@mysign.ch wrote:
 Given the following directory/file structure

 .../solrhome
 .../solrhome/solr.xml
 .../solrhome/config/
 .../solrhome/config/schema.xml
 .../solrhome/config/solrconfig.xml
 .../solrhome/cores

 How can I programmatically (CoreContainer) create a new core below 
 .../solrhome/cores that makes use of solrconfig.xml and schema.xml?
 The new core should use the config files in .../solrhome/config/

 Thx
 Clemens


Re: Backuping SolrCloud

2014-11-24 Thread Vivek Pathak
I was able to get very good backup procedure by having cron job perform compact 
on each shard and then copy out the physical shard (the full lucene index 
directory to a backup server)

Updates would need to be stopped during this period.   And restore would be as 
simple as copying in the backed up shard and restarting solr 

 On Nov 24, 2014, at 9:48 AM, elmerfudd na...@012.net.il wrote:
 
 Hi, I'm looking for a built-in SolrCloud backup mechanism.
 I want to backup my Index (scheduled / manual backups) while Indexing and
 searching.
 
 What is the proper way to perform this backup-restore task?
 
 
 Thanks.
 
 
 
 
 --
 View this message in context: 
 http://lucene.472066.n3.nabble.com/Backuping-SolrCloud-tp4170624.html
 Sent from the Solr - User mailing list archive at Nabble.com.


Re: CoreContainer : create new cores reusing/sharing solrconfig.xml and schema.xml

2014-11-24 Thread Alexandre Rafalovitch
I haven't done this myself, but it is defined as a constant in the
CoreDescriptor:
http://www.solr-start.com/javadoc/solr-lucene/org/apache/solr/core/CoreDescriptor.html#CORE_CONFIGSET

You could also look at the tests and see whether it is used.

Regards,
   Alex.
Personal: http://www.outerthoughts.com/ and @arafalov
Solr resources and newsletter: http://www.solr-start.com/ and @solrstart
Solr popularizers community: https://www.linkedin.com/groups?gid=6713853


On 24 November 2014 at 10:43, Clemens Wyss DEV clemens...@mysign.ch wrote:
 Thanks fort he quick reply.
 But how do I hand in the configSet to use. I guess through the CoreDescriptor?

 -Ursprüngliche Nachricht-
 Von: Alexandre Rafalovitch [mailto:arafa...@gmail.com]
 Gesendet: Montag, 24. November 2014 16:27
 An: solr-user
 Betreff: Re: CoreContainer : create new cores reusing/sharing solrconfig.xml 
 and schema.xml

 Have you tried Configsets:
 https://cwiki.apache.org/confluence/display/solr/Config+Sets ?

 Regards,
Alex.
 Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and 
 newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers 
 community: https://www.linkedin.com/groups?gid=6713853


 On 24 November 2014 at 10:14, Clemens Wyss DEV clemens...@mysign.ch wrote:
 Given the following directory/file structure

 .../solrhome
 .../solrhome/solr.xml
 .../solrhome/config/
 .../solrhome/config/schema.xml
 .../solrhome/config/solrconfig.xml
 .../solrhome/cores

 How can I programmatically (CoreContainer) create a new core below 
 .../solrhome/cores that makes use of solrconfig.xml and schema.xml?
 The new core should use the config files in .../solrhome/config/

 Thx
 Clemens


AW: CoreContainer : create new cores reusing/sharing solrconfig.xml and schema.xml

2014-11-24 Thread Clemens Wyss DEV
Looks like 4.7.2 does not have this const. Another reason to switch to 4.10 
(and java 8 ;) )

-Ursprüngliche Nachricht-
Von: Alexandre Rafalovitch [mailto:arafa...@gmail.com] 
Gesendet: Montag, 24. November 2014 16:49
An: solr-user
Betreff: Re: CoreContainer : create new cores reusing/sharing solrconfig.xml 
and schema.xml

I haven't done this myself, but it is defined as a constant in the
CoreDescriptor:
http://www.solr-start.com/javadoc/solr-lucene/org/apache/solr/core/CoreDescriptor.html#CORE_CONFIGSET

You could also look at the tests and see whether it is used.

Regards,
   Alex.
Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and 
newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers 
community: https://www.linkedin.com/groups?gid=6713853


On 24 November 2014 at 10:43, Clemens Wyss DEV clemens...@mysign.ch wrote:
 Thanks fort he quick reply.
 But how do I hand in the configSet to use. I guess through the CoreDescriptor?

 -Ursprüngliche Nachricht-
 Von: Alexandre Rafalovitch [mailto:arafa...@gmail.com]
 Gesendet: Montag, 24. November 2014 16:27
 An: solr-user
 Betreff: Re: CoreContainer : create new cores reusing/sharing 
 solrconfig.xml and schema.xml

 Have you tried Configsets:
 https://cwiki.apache.org/confluence/display/solr/Config+Sets ?

 Regards,
Alex.
 Personal: http://www.outerthoughts.com/ and @arafalov Solr resources 
 and newsletter: http://www.solr-start.com/ and @solrstart Solr 
 popularizers community: https://www.linkedin.com/groups?gid=6713853


 On 24 November 2014 at 10:14, Clemens Wyss DEV clemens...@mysign.ch wrote:
 Given the following directory/file structure

 .../solrhome
 .../solrhome/solr.xml
 .../solrhome/config/
 .../solrhome/config/schema.xml
 .../solrhome/config/solrconfig.xml
 .../solrhome/cores

 How can I programmatically (CoreContainer) create a new core below 
 .../solrhome/cores that makes use of solrconfig.xml and schema.xml?
 The new core should use the config files in .../solrhome/config/

 Thx
 Clemens


Re: Backuping SolrCloud

2014-11-24 Thread Erick Erickson
There are some JIRAs out there for better backup/restore for SolrCloud
situations, but
nothing has been committed yet that I know of.

Best,
Erick

On Mon, Nov 24, 2014 at 7:45 AM, Vivek Pathak vpat...@orgmeta.com wrote:
 I was able to get very good backup procedure by having cron job perform 
 compact on each shard and then copy out the physical shard (the full lucene 
 index directory to a backup server)

 Updates would need to be stopped during this period.   And restore would be 
 as simple as copying in the backed up shard and restarting solr

 On Nov 24, 2014, at 9:48 AM, elmerfudd na...@012.net.il wrote:

 Hi, I'm looking for a built-in SolrCloud backup mechanism.
 I want to backup my Index (scheduled / manual backups) while Indexing and
 searching.

 What is the proper way to perform this backup-restore task?


 Thanks.




 --
 View this message in context: 
 http://lucene.472066.n3.nabble.com/Backuping-SolrCloud-tp4170624.html
 Sent from the Solr - User mailing list archive at Nabble.com.


Re: Backuping SolrCloud

2014-11-24 Thread ralph tice
I have a writeup of how to perform safe backups here:
https://gist.github.com/ralph-tice/887414a7f8082a0cb828

There are some tickets around this work to further the ease of
backups, especially https://issues.apache.org/jira/browse/SOLR-5750

On Mon, Nov 24, 2014 at 9:45 AM, Vivek Pathak vpat...@orgmeta.com wrote:
 I was able to get very good backup procedure by having cron job perform 
 compact on each shard and then copy out the physical shard (the full lucene 
 index directory to a backup server)

 Updates would need to be stopped during this period.   And restore would be 
 as simple as copying in the backed up shard and restarting solr

 On Nov 24, 2014, at 9:48 AM, elmerfudd na...@012.net.il wrote:

 Hi, I'm looking for a built-in SolrCloud backup mechanism.
 I want to backup my Index (scheduled / manual backups) while Indexing and
 searching.

 What is the proper way to perform this backup-restore task?


 Thanks.




 --
 View this message in context: 
 http://lucene.472066.n3.nabble.com/Backuping-SolrCloud-tp4170624.html
 Sent from the Solr - User mailing list archive at Nabble.com.


Re: CoreContainer : create new cores reusing/sharing solrconfig.xml and schema.xml

2014-11-24 Thread Erick Erickson
4.10 does NOT require java 8, it requires Java 1.7. In fact, 5.x is on
Java 1.7 too.

Trunk, (the future Solr/Lucene 6.0) is on Java 8 though (just recently moved).

But if you wan to go to Java 8 anyway and need an excuse pretend I didn't write
this ;).

Best,
Erick

On Mon, Nov 24, 2014 at 7:58 AM, Clemens Wyss DEV clemens...@mysign.ch wrote:
 Looks like 4.7.2 does not have this const. Another reason to switch to 4.10 
 (and java 8 ;) )

 -Ursprüngliche Nachricht-
 Von: Alexandre Rafalovitch [mailto:arafa...@gmail.com]
 Gesendet: Montag, 24. November 2014 16:49
 An: solr-user
 Betreff: Re: CoreContainer : create new cores reusing/sharing solrconfig.xml 
 and schema.xml

 I haven't done this myself, but it is defined as a constant in the
 CoreDescriptor:
 http://www.solr-start.com/javadoc/solr-lucene/org/apache/solr/core/CoreDescriptor.html#CORE_CONFIGSET

 You could also look at the tests and see whether it is used.

 Regards,
Alex.
 Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and 
 newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers 
 community: https://www.linkedin.com/groups?gid=6713853


 On 24 November 2014 at 10:43, Clemens Wyss DEV clemens...@mysign.ch wrote:
 Thanks fort he quick reply.
 But how do I hand in the configSet to use. I guess through the 
 CoreDescriptor?

 -Ursprüngliche Nachricht-
 Von: Alexandre Rafalovitch [mailto:arafa...@gmail.com]
 Gesendet: Montag, 24. November 2014 16:27
 An: solr-user
 Betreff: Re: CoreContainer : create new cores reusing/sharing
 solrconfig.xml and schema.xml

 Have you tried Configsets:
 https://cwiki.apache.org/confluence/display/solr/Config+Sets ?

 Regards,
Alex.
 Personal: http://www.outerthoughts.com/ and @arafalov Solr resources
 and newsletter: http://www.solr-start.com/ and @solrstart Solr
 popularizers community: https://www.linkedin.com/groups?gid=6713853


 On 24 November 2014 at 10:14, Clemens Wyss DEV clemens...@mysign.ch wrote:
 Given the following directory/file structure

 .../solrhome
 .../solrhome/solr.xml
 .../solrhome/config/
 .../solrhome/config/schema.xml
 .../solrhome/config/solrconfig.xml
 .../solrhome/cores

 How can I programmatically (CoreContainer) create a new core below 
 .../solrhome/cores that makes use of solrconfig.xml and schema.xml?
 The new core should use the config files in .../solrhome/config/

 Thx
 Clemens


Re: use {!surround} with more conditions

2014-11-24 Thread Ahmet Arslan
Hi,

Not sure about SomeField:* but others should work.

You can move clauses that do not require surround to filter queries.

e.g. fq=-SomeField:Text3*

Ahmet


On Monday, November 24, 2014 4:33 PM, Bet Hel beteli8...@gmail.com wrote:
Can't make surround work with other conditions...
for example:

{!surround}SomeField:30w(Text1*,Text2*)  -SomeField:Text3*

SomeField:* -{!surround}SomeField:30w(Text1*,Text2*)

Is that doable?


Re: how solr4.0 and zookeeper run on weblogic

2014-11-24 Thread Ramzi Alqrainy
I hope this post helps you

http://lucene.472066.n3.nabble.com/how-solr4-0-and-zookeeper-run-on-weblogic-td4013882.html



--
View this message in context: 
http://lucene.472066.n3.nabble.com/how-solr4-0-and-zookeeper-run-on-weblogic-tp4170625p4170679.html
Sent from the Solr - User mailing list archive at Nabble.com.


Documents to query

2014-11-24 Thread Robust Links
Hi

Is there a way (in Lucene / SOLR) to go from a vector of documents (IDs,
and optionally their scores) to a query? i.e. the reverse process of query
- document?

thank you

Peyman


Re: Documents to query

2014-11-24 Thread Alexandre Rafalovitch
Very unlikely. What's the business case?

Regards,
   Alex.
Personal: http://www.outerthoughts.com/ and @arafalov
Solr resources and newsletter: http://www.solr-start.com/ and @solrstart
Solr popularizers community: https://www.linkedin.com/groups?gid=6713853


On 24 November 2014 at 13:23, Robust Links pey...@robustlinks.com wrote:
 Hi

 Is there a way (in Lucene / SOLR) to go from a vector of documents (IDs,
 and optionally their scores) to a query? i.e. the reverse process of query
 - document?

 thank you

 Peyman


Re: Documents to query

2014-11-24 Thread Walter Underwood
This feature is called “more like this”. I think it only works for a single 
document, but it probably could be extended.

wunder
Walter Underwood
wun...@wunderwood.org
http://observer.wunderwood.org/

On Nov 24, 2014, at 10:26 AM, Alexandre Rafalovitch arafa...@gmail.com wrote:

 Very unlikely. What's the business case?
 
 Regards,
   Alex.
 Personal: http://www.outerthoughts.com/ and @arafalov
 Solr resources and newsletter: http://www.solr-start.com/ and @solrstart
 Solr popularizers community: https://www.linkedin.com/groups?gid=6713853
 
 
 On 24 November 2014 at 13:23, Robust Links pey...@robustlinks.com wrote:
 Hi
 
 Is there a way (in Lucene / SOLR) to go from a vector of documents (IDs,
 and optionally their scores) to a query? i.e. the reverse process of query
 - document?
 
 thank you
 
 Peyman



matching shingles issue

2014-11-24 Thread vit
I have Solr 4.2.1
I am using the following analyser:
fieldType name=text_shingle class=solr.TextField
positionIncrementGap=100
analyzer type=index
tokenizer class=solr.WhitespaceTokenizerFactory/
filter class=solr.ShingleFilterFactory 
minShingleSize=2
maxShingleSize=5
outputUnigrams=true 
outputUnigramsIfNoShingles=false
tokenSeparator= /
/analyzer
analyzer type=query
tokenizer class=solr.WhitespaceTokenizerFactory/
filter class=solr.ShingleFilterFactory 
minShingleSize=2
maxShingleSize=5
outputUnigrams=false 
outputUnigramsIfNoShingles=true
tokenSeparator= /
/analyzer
/fieldType



for Query: 
description_shingle:Highest quality

I am getting Result:
arr name=description_shingle
  strHighest standards of quality installations!/str
/arr

So the result does not have shingle Highest quality
Instead it has 
Highest standards of quality

The question is why I am getting this match



--
View this message in context: 
http://lucene.472066.n3.nabble.com/matching-shingles-issue-tp4170685.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Documents to query

2014-11-24 Thread Robust Links
@ Alexandre.. the business case is the following. I am using lucene/solr to
compose vectors. I want to apply different functions (addition,
multipication, tensor products,etc) to these vectors. The end result of
that is a new vector. I would like to then go back from that vector back to
term that might have generated that vector..i.e. the following pipeline

1) term i - query i - N docs_of_i
2) term j - query j - N docs_of_j
3) F(docs_of_i, docs_of_j) - N docs
4) N docs - term_k

1  2 are usual query path. steps 3 and 4 are what i am trying to do

@walter. if I am correct, MLT will generate a term for *EACH* element of
the vector (N docs above).. but I am interested in a single term



On Mon, Nov 24, 2014 at 1:31 PM, Walter Underwood wun...@wunderwood.org
wrote:

 This feature is called “more like this”. I think it only works for a
 single document, but it probably could be extended.

 wunder
 Walter Underwood
 wun...@wunderwood.org
 http://observer.wunderwood.org/

 On Nov 24, 2014, at 10:26 AM, Alexandre Rafalovitch arafa...@gmail.com
 wrote:

  Very unlikely. What's the business case?
 
  Regards,
Alex.
  Personal: http://www.outerthoughts.com/ and @arafalov
  Solr resources and newsletter: http://www.solr-start.com/ and @solrstart
  Solr popularizers community: https://www.linkedin.com/groups?gid=6713853
 
 
  On 24 November 2014 at 13:23, Robust Links pey...@robustlinks.com
 wrote:
  Hi
 
  Is there a way (in Lucene / SOLR) to go from a vector of documents (IDs,
  and optionally their scores) to a query? i.e. the reverse process of
 query
  - document?
 
  thank you
 
  Peyman




Re: Indexing with SolrJ fails on windows

2014-11-24 Thread brian4
The problem seems to occur at the apache redirect - I found if I bypass
apache by using my VM IP address directly as the Solr URL, then the error
does not occur even from windows.

From some searching it seems like Apache does not allow carriage returns in
its request headers - so my guess is SolrJ is adding a carriage return in
the request when run on Windows, but not on linux, so when receiving the
request from Windows apache is spitting back an error.

Is there any way to disable this behavior with SolrJ / have it generate
consistent requests regardless of platform?



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Indexing-with-SolrJ-fails-on-windows-tp4170687p4170690.html
Sent from the Solr - User mailing list archive at Nabble.com.


Indexing with SolrJ fails on windows

2014-11-24 Thread brian4
I am using solrj to index to Solr through a Java application - I've tried
this both with Solr 4.8.1 and Solr 4.10.2, indexing to Solr 4.10.0.

I've found I cannot index too large content (a field with 400 words) or more
than 1 document at once to Solr instances from windows.  The exact same
indexing code works from linux, unchanged.  

I've found it does not work on Windows if:
1) I try to add more than 1 document at a time
2) I try to add a document with a long field value (400 words).

However in both cases it works fine if run from linux, or on windows if I
only add one document without very long values for any field.

The exception I get is the following:
org.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException:
Invalid chunk header
at
org.apache.solr.client.solrj.impl.HttpSolrServer.executeMethod(HttpSolrServer.java:552)
~[solr-solrj-4.10.2.jar:4.10.2 1634293 - mike - 2014-10-26 05:56:22]
at
org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:210)
~[solr-solrj-4.10.2.jar:4.10.2 1634293 - mike - 2014-10-26 05:56:22]
at
org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:206)
~[solr-solrj-4.10.2.jar:4.10.2 1634293 - mike - 2014-10-26 05:56:22]
at
org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:124)
~[solr-solrj-4.10.2.jar:4.10.2 1634293 - mike - 2014-10-26 05:56:22]
at org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:68)
~[solr-solrj-4.10.2.jar:4.10.2 1634293 - mike - 2014-10-26 05:56:22]
at org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:54)
~[solr-solrj-4.10.2.jar:4.10.2 1634293 - mike - 2014-10-26 05:56:22]


My test code to reproduce this is the following:
@Test
public void testWriteBigOther() throws Exception {
SolrServer solrServer = new
HttpSolrServer(http://my-vm:8080/solr/test_copy;);

SolrInputDocument doc = new SolrInputDocument();
doc.addField(asset_id,test_write_big);
doc.addField(title, test write big title); 
doc.addField(secondary_header, StringUtils.repeat(396, ,400)); 

ListSolrInputDocument inputDocs = new
ArrayListSolrInputDocument();
inputDocs.add(doc);

solrServer.add(inputDocs);

solrServer.commit();
solrServer.shutdown();
}

(it uses org.apache.commons.lang3.StringUtils repeat() method to generate
the large field value).


It seems like there must be a bug in SolrJ - i.e., I guess when it is
building the request it does something differently in windows vs. linux -
like maybe adds a carriage return on windows?  Does anyone know how to fix
this, or what else I could do to diagnose it?




--
View this message in context: 
http://lucene.472066.n3.nabble.com/Indexing-with-SolrJ-fails-on-windows-tp4170687.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Documents to query

2014-11-24 Thread Alexandre Rafalovitch
Is this the Paris - France + Italy = Rome thing? I think I've seen a
couple of discussions about that.

When you say a term are you actually looking for one keyword?
Because that's very different from a query that can be infinitely
complex.

Regards,
   Alex.
Personal: http://www.outerthoughts.com/ and @arafalov
Solr resources and newsletter: http://www.solr-start.com/ and @solrstart
Solr popularizers community: https://www.linkedin.com/groups?gid=6713853


On 24 November 2014 at 13:44, Robust Links pey...@robustlinks.com wrote:
 @ Alexandre.. the business case is the following. I am using lucene/solr to
 compose vectors. I want to apply different functions (addition,
 multipication, tensor products,etc) to these vectors. The end result of
 that is a new vector. I would like to then go back from that vector back to
 term that might have generated that vector..i.e. the following pipeline

 1) term i - query i - N docs_of_i
 2) term j - query j - N docs_of_j
 3) F(docs_of_i, docs_of_j) - N docs
 4) N docs - term_k

 1  2 are usual query path. steps 3 and 4 are what i am trying to do

 @walter. if I am correct, MLT will generate a term for *EACH* element of
 the vector (N docs above).. but I am interested in a single term



 On Mon, Nov 24, 2014 at 1:31 PM, Walter Underwood wun...@wunderwood.org
 wrote:

 This feature is called “more like this”. I think it only works for a
 single document, but it probably could be extended.

 wunder
 Walter Underwood
 wun...@wunderwood.org
 http://observer.wunderwood.org/

 On Nov 24, 2014, at 10:26 AM, Alexandre Rafalovitch arafa...@gmail.com
 wrote:

  Very unlikely. What's the business case?
 
  Regards,
Alex.
  Personal: http://www.outerthoughts.com/ and @arafalov
  Solr resources and newsletter: http://www.solr-start.com/ and @solrstart
  Solr popularizers community: https://www.linkedin.com/groups?gid=6713853
 
 
  On 24 November 2014 at 13:23, Robust Links pey...@robustlinks.com
 wrote:
  Hi
 
  Is there a way (in Lucene / SOLR) to go from a vector of documents (IDs,
  and optionally their scores) to a query? i.e. the reverse process of
 query
  - document?
 
  thank you
 
  Peyman




Re: matching shingles issue

2014-11-24 Thread Michael Sokolov

maybe try

description_shingle:(Highest quality)


On 11/24/14 1:46 PM, vit wrote:

I have Solr 4.2.1
I am using the following analyser:
fieldType name=text_shingle class=solr.TextField
positionIncrementGap=100
analyzer type=index
tokenizer class=solr.WhitespaceTokenizerFactory/
filter class=solr.ShingleFilterFactory 
minShingleSize=2
maxShingleSize=5
outputUnigrams=true 
outputUnigramsIfNoShingles=false
tokenSeparator= /
/analyzer
analyzer type=query
tokenizer class=solr.WhitespaceTokenizerFactory/
filter class=solr.ShingleFilterFactory 
minShingleSize=2
maxShingleSize=5
outputUnigrams=false 
outputUnigramsIfNoShingles=true
tokenSeparator= /
/analyzer
/fieldType



for Query:
description_shingle:Highest quality

I am getting Result:
arr name=description_shingle
   strHighest standards of quality installations!/str
/arr

So the result does not have shingle Highest quality
Instead it has
Highest standards of quality

The question is why I am getting this match



--
View this message in context: 
http://lucene.472066.n3.nabble.com/matching-shingles-issue-tp4170685.html
Sent from the Solr - User mailing list archive at Nabble.com.




Re: Indexing with SolrJ fails on windows

2014-11-24 Thread Alexandre Rafalovitch
Try running the client with -Dline.separator='\n' to force the line
separator. 
https://docs.oracle.com/javase/tutorial/essential/environment/sysprop.html

However, if it that does work, it's probably a bug.

Regards,
   Alex.
Personal: http://www.outerthoughts.com/ and @arafalov
Solr resources and newsletter: http://www.solr-start.com/ and @solrstart
Solr popularizers community: https://www.linkedin.com/groups?gid=6713853


On 24 November 2014 at 14:00, brian4 bqu...@gmail.com wrote:
 The problem seems to occur at the apache redirect - I found if I bypass
 apache by using my VM IP address directly as the Solr URL, then the error
 does not occur even from windows.

 From some searching it seems like Apache does not allow carriage returns in
 its request headers - so my guess is SolrJ is adding a carriage return in
 the request when run on Windows, but not on linux, so when receiving the
 request from Windows apache is spitting back an error.

 Is there any way to disable this behavior with SolrJ / have it generate
 consistent requests regardless of platform?



 --
 View this message in context: 
 http://lucene.472066.n3.nabble.com/Indexing-with-SolrJ-fails-on-windows-tp4170687p4170690.html
 Sent from the Solr - User mailing list archive at Nabble.com.


Re: Documents to query

2014-11-24 Thread Robust Links
i term is the title of the document indexed. So its not an unbounded
query term.

On Mon, Nov 24, 2014 at 2:02 PM, Alexandre Rafalovitch arafa...@gmail.com
wrote:

 Is this the Paris - France + Italy = Rome thing? I think I've seen a
 couple of discussions about that.

 When you say a term are you actually looking for one keyword?
 Because that's very different from a query that can be infinitely
 complex.

 Regards,
Alex.
 Personal: http://www.outerthoughts.com/ and @arafalov
 Solr resources and newsletter: http://www.solr-start.com/ and @solrstart
 Solr popularizers community: https://www.linkedin.com/groups?gid=6713853


 On 24 November 2014 at 13:44, Robust Links pey...@robustlinks.com wrote:
  @ Alexandre.. the business case is the following. I am using lucene/solr
 to
  compose vectors. I want to apply different functions (addition,
  multipication, tensor products,etc) to these vectors. The end result of
  that is a new vector. I would like to then go back from that vector back
 to
  term that might have generated that vector..i.e. the following pipeline
 
  1) term i - query i - N docs_of_i
  2) term j - query j - N docs_of_j
  3) F(docs_of_i, docs_of_j) - N docs
  4) N docs - term_k
 
  1  2 are usual query path. steps 3 and 4 are what i am trying to do
 
  @walter. if I am correct, MLT will generate a term for *EACH* element of
  the vector (N docs above).. but I am interested in a single term
 
 
 
  On Mon, Nov 24, 2014 at 1:31 PM, Walter Underwood wun...@wunderwood.org
 
  wrote:
 
  This feature is called “more like this”. I think it only works for a
  single document, but it probably could be extended.
 
  wunder
  Walter Underwood
  wun...@wunderwood.org
  http://observer.wunderwood.org/
 
  On Nov 24, 2014, at 10:26 AM, Alexandre Rafalovitch arafa...@gmail.com
 
  wrote:
 
   Very unlikely. What's the business case?
  
   Regards,
 Alex.
   Personal: http://www.outerthoughts.com/ and @arafalov
   Solr resources and newsletter: http://www.solr-start.com/ and
 @solrstart
   Solr popularizers community:
 https://www.linkedin.com/groups?gid=6713853
  
  
   On 24 November 2014 at 13:23, Robust Links pey...@robustlinks.com
  wrote:
   Hi
  
   Is there a way (in Lucene / SOLR) to go from a vector of documents
 (IDs,
   and optionally their scores) to a query? i.e. the reverse process of
  query
   - document?
  
   thank you
  
   Peyman
 
 



Help on Notification for SOLR

2014-11-24 Thread EXTERNAL Taminidi Ravi (ETI, Automotive-Service-Solutions)
Hello,  

I am working on implementing a notification system when there is any change in 
the solr index (Addition or removal of documents from Index).

Can someone guide me on the starting point. I see some articles using DIH? Is 
there any way using without DIH.

Thanks

Ravi


Re: Documents to query

2014-11-24 Thread Mikhail Khludnev
Doesn't it sound like faceting?

On Mon, Nov 24, 2014 at 9:44 PM, Robust Links pey...@robustlinks.com
wrote:

 @ Alexandre.. the business case is the following. I am using lucene/solr to
 compose vectors. I want to apply different functions (addition,
 multipication, tensor products,etc) to these vectors. The end result of
 that is a new vector. I would like to then go back from that vector back to
 term that might have generated that vector..i.e. the following pipeline

 1) term i - query i - N docs_of_i
 2) term j - query j - N docs_of_j
 3) F(docs_of_i, docs_of_j) - N docs
 4) N docs - term_k

 1  2 are usual query path. steps 3 and 4 are what i am trying to do

 @walter. if I am correct, MLT will generate a term for *EACH* element of
 the vector (N docs above).. but I am interested in a single term



 On Mon, Nov 24, 2014 at 1:31 PM, Walter Underwood wun...@wunderwood.org
 wrote:

  This feature is called “more like this”. I think it only works for a
  single document, but it probably could be extended.
 
  wunder
  Walter Underwood
  wun...@wunderwood.org
  http://observer.wunderwood.org/
 
  On Nov 24, 2014, at 10:26 AM, Alexandre Rafalovitch arafa...@gmail.com
  wrote:
 
   Very unlikely. What's the business case?
  
   Regards,
 Alex.
   Personal: http://www.outerthoughts.com/ and @arafalov
   Solr resources and newsletter: http://www.solr-start.com/ and
 @solrstart
   Solr popularizers community:
 https://www.linkedin.com/groups?gid=6713853
  
  
   On 24 November 2014 at 13:23, Robust Links pey...@robustlinks.com
  wrote:
   Hi
  
   Is there a way (in Lucene / SOLR) to go from a vector of documents
 (IDs,
   and optionally their scores) to a query? i.e. the reverse process of
  query
   - document?
  
   thank you
  
   Peyman
 
 




-- 
Sincerely yours
Mikhail Khludnev
Principal Engineer,
Grid Dynamics

http://www.griddynamics.com
mkhlud...@griddynamics.com


RE: Help on Notification for SOLR

2014-11-24 Thread Gonzalo Rodriguez
Hi,

You might want to take a look at the UpdateRequestProcessor and 
UpdateRequestProcessorChain classes:

https://wiki.apache.org/solr/UpdateRequestProcessor and 
http://lucene.apache.org/solr/4_10_2/solr-core/org/apache/solr/update/processor/UpdateRequestProcessorChain.html

You can hook a chain to your update request handler and use a 
ScriptUpdateProcessor to implement your notification system.

https://wiki.apache.org/solr/ScriptUpdateProcessor

Hope this helps,
Gonzalo

-Original Message-
From: EXTERNAL Taminidi Ravi (ETI, Automotive-Service-Solutions) 
[mailto:external.ravi.tamin...@us.bosch.com] 
Sent: Monday, November 24, 2014 2:35 PM
To: solr-user@lucene.apache.org
Subject: Help on Notification for SOLR

Hello,  

I am working on implementing a notification system when there is any change in 
the solr index (Addition or removal of documents from Index).

Can someone guide me on the starting point. I see some articles using DIH? Is 
there any way using without DIH.

Thanks

Ravi


Re: Help on Notification for SOLR

2014-11-24 Thread Alexandre Rafalovitch
There are commit notifications scripts in solrconfig.xml, there is JMX,
there is whatever the admin UI uses for freeze and show change screen.
And, as mentioned already, URP chains.

Regards,
 Alex
On 24/11/2014 3:36 pm, EXTERNAL Taminidi Ravi (ETI,
Automotive-Service-Solutions) external.ravi.tamin...@us.bosch.com wrote:

 Hello,

 I am working on implementing a notification system when there is any
 change in the solr index (Addition or removal of documents from Index).

 Can someone guide me on the starting point. I see some articles using DIH?
 Is there any way using without DIH.

 Thanks

 Ravi



Re: Help on Notification for SOLR

2014-11-24 Thread Ramzi Alqrainy
of course there is another approach which is using Solr Clients. So, you can
use solr clients for specific language like PHP, JAVA, ...

For Example :

JAVA Solr Client :

SolrJ is an API that makes it easy for Java applications to talk to Solr.
SolrJ hides a lot of the details of connecting to Solr and allows your
application to interact with Solr with simple high-level methods.


String urlString = http://localhost:8983/solr;;
SolrServer solr = new HttpSolrServer(urlString);
SolrInputDocument document = new SolrInputDocument();
document.addField(id, 552199);
document.addField(name, Gouda cheese wheel);
document.addField(price, 49.99);
UpdateResponse response = solr.add(document);
 
// Remember to commit your changes!
 
solr.commit();


If you want another language, you can tell me and I will help you .


   
EXTERNAL Taminidi Ravi (ETI, Automotive-Service-Solutions) wrote
 Hello,  
 
 I am working on implementing a notification system when there is any
 change in the solr index (Addition or removal of documents from Index).
 
 Can someone guide me on the starting point. I see some articles using DIH?
 Is there any way using without DIH.
 
 Thanks
 
 Ravi





--
View this message in context: 
http://lucene.472066.n3.nabble.com/Help-on-Notification-for-SOLR-tp4170708p4170716.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Using stored value of a field to build suggester index

2014-11-24 Thread Faisal Mansoor
Thanks Erick, This makes things clearer.

Thanks,
Faisal

On Sun, Nov 23, 2014 at 2:17 PM, Erick Erickson erickerick...@gmail.com
wrote:

 You can't build the suggester from the stored
 values, it's constructed from indexed terms only.

 You probably want to create a copyField to
 a less-analyzed (indexed) field and suggest
 from _that_. You'll probably want to do things
 like remove punctuation, perhaps lowercase
 and the like but not stem etc.

 Best,
 Erick

 On Sun, Nov 23, 2014 at 12:25 PM, Faisal Mansoor
 faisal.mans...@gmail.com wrote:
  Hi,
 
  I am trying to build a suggester for a field which is both index and
  stored. The field is whitespace tokenized, lowercased, stemmed etc while
  indexing.
 
  It looks like that the indexed terms are used as a source for building
 the
  suggester index. Which is what the following line in the suggester
  documentation also mentions.
 
  https://wiki.apache.org/solr/Suggester
 
 
 -
 
 field - if sourceLocation is empty then terms from this field in the
 index will be used when building the trie.
 
  I want to display the suggested value in UI, is it possible to use the
  stored value of the field rather than the indexed terms to build the
 index.
 
  Here are the relevant definitions from solrconfig.xml and schema.xml.
 
  Thanks.
  Faisal
 
  solrconfig.xml
 
searchComponent class=solr.SpellCheckComponent
  name=infix_suggest_analyzing
  lst name=spellchecker
str name=nameinfix_suggest_analyzing/str
str
 name=classnameorg.apache.solr.spelling.suggest.Suggester/str
str
 
 name=lookupImplorg.apache.solr.spelling.suggest.fst.AnalyzingInfixLookupFactory/str
str name=buildOnCommitfalse/str
 
!-- Suggester properties --
str name=suggestAnalyzerFieldTypeautosuggest_fieldType/str
str
 
 name=dictionaryImplorg.apache.solr.spelling.suggest.HighFrequencyDictionaryFactory/str
str name=fieldDisplayName/str
 
 
  /lst
 
  !-- specify a fieldtype using keywordtokenizer + lowercase + cleanup
  --
  str name=queryAnalyzerFieldTypephrase_suggest/str
/searchComponent
 
 
requestHandler name=/suggest
  class=org.apache.solr.handler.component.SearchHandler
  lst name=defaults
str name=echoParamsexplicit/str
str name=spellchecktrue/str
str name=spellcheck.dictionaryinfix_suggest_analyzing/str
str name=spellcheck.onlyMorePopulartrue/str
str name=spellcheck.count200/str
str name=spellcheck.collatetrue/str
str name=spellcheck.maxCollations10/str
  /lst
  arr name=components
strinfix_suggest_analyzing/str
  /arr
/requestHandler
 
 
  schema.xml
 
 
fieldType name=autosuggest_fieldType class=solr.TextField
  positionIncrementGap=100
  analyzer
tokenizer class=solr.StandardTokenizerFactory/
filter class=solr.LowerCaseFilterFactory/
filter class=solr.ASCIIFoldingFilterFactory/
  /analyzer
/fieldType
 
fieldtype name=phrase_suggest class=solr.TextField
analyzer
  tokenizer class=solr.KeywordTokenizerFactory/
  filter class=solr.PatternReplaceFilterFactory
 
 
 pattern=([^\p{L}\p{M}\p{N}\p{Cs}]*[\p{L}\p{M}\p{N}\p{Cs}\_]+:)|([^\p{L}\p{M}\p{N}\p{Cs}])+
  replacement=  replace=all/
  filter class=solr.LowerCaseFilterFactory/
  filter class=solr.TrimFilterFactory/
/analyzer
  /fieldtype
 
 
  fieldType name=text class=solr.TextField
  positionIncrementGap=100
analyzer type=index
  tokenizer class=solr.WhitespaceTokenizerFactory/
  filter class=solr.StopFilterFactory ignoreCase=true
  words=stopwords.txt/
  filter class=solr.WordDelimiterFilterFactory
  generateWordParts=1 generateNumberParts=1 catenateWords=1
  catenateNumbers=1 catenateAll=0 splitOnCaseChange=1/
  filter class=solr.LowerCaseFilterFactory/
  filter class=solr.KeywordMarkerFilterFactory
  protected=protwords.txt/
  filter class=solr.PorterStemFilterFactory/
  filter class=solr.RemoveDuplicatesTokenFilterFactory/
/analyzer
analyzer type=query
  tokenizer class=solr.WhitespaceTokenizerFactory/
  filter class=solr.SynonymFilterFactory synonyms=synonyms.txt
  ignoreCase=true expand=true/
  filter class=solr.StopFilterFactory ignoreCase=true
  words=stopwords.txt/
  filter class=solr.WordDelimiterFilterFactory
  generateWordParts=1 generateNumberParts=1 catenateWords=0
  catenateNumbers=0 catenateAll=0 splitOnCaseChange=1/
  filter class=solr.LowerCaseFilterFactory/
  filter class=solr.KeywordMarkerFilterFactory
  protected=protwords.txt/
  filter class=solr.PorterStemFilterFactory/
  filter class=solr.RemoveDuplicatesTokenFilterFactory/
/analyzer
  /fieldType
 
 
  field name=DisplayName type=text indexed=true stored=true
  required=true multiValued=false /



solrj getBeans from GroupCommand

2014-11-24 Thread Phil Bartholomew
I'm fairly new to Solr but think I have read the documentation well.

Really enjoying the ease of using the likes of:

ArrayListMyClass myRecords = queryResponse.getBeans(MyClass.class);

However I have now come across the situation where I need to use groups. I
cannot seem to find a way to getBeans() from a group, maybe I'm not
understanding the hierarchy within a group? But I was hoping for something
along the lines of:

//Get beans from each group returnedList GroupCommand groupCommands
= queryResponse.getGroupResponse().getValues();for(GroupCommand gc :
groupCommands){
ListGroup groups = gc.getValues();
Group group = groups.get(0);
ArrayListMyClass myRecords = group.getBeans(MyClass.class);
//Do something with the current group of beans
}

Any advice or comments on how to do this or an alternative method would be
gratefully received.

Many thanks

I have also posted this at:
http://stackoverflow.com/questions/27114168/solrj-getbeans-from-groupcommand


Re: Backuping SolrCloud

2014-11-24 Thread Noble Paul
Hi,
There is a ticket for the same .

https://issues.apache.org/jira/browse/SOLR-5750

Would you mind contributing to the discussion there?


On Mon, Nov 24, 2014 at 11:07 AM, ralph tice ralph.t...@gmail.com wrote:

 I have a writeup of how to perform safe backups here:
 https://gist.github.com/ralph-tice/887414a7f8082a0cb828

 There are some tickets around this work to further the ease of
 backups, especially https://issues.apache.org/jira/browse/SOLR-5750

 On Mon, Nov 24, 2014 at 9:45 AM, Vivek Pathak vpat...@orgmeta.com wrote:
  I was able to get very good backup procedure by having cron job perform
 compact on each shard and then copy out the physical shard (the full lucene
 index directory to a backup server)
 
  Updates would need to be stopped during this period.   And restore would
 be as simple as copying in the backed up shard and restarting solr
 
  On Nov 24, 2014, at 9:48 AM, elmerfudd na...@012.net.il wrote:
 
  Hi, I'm looking for a built-in SolrCloud backup mechanism.
  I want to backup my Index (scheduled / manual backups) while Indexing
 and
  searching.
 
  What is the proper way to perform this backup-restore task?
 
 
  Thanks.
 
 
 
 
  --
  View this message in context:
 http://lucene.472066.n3.nabble.com/Backuping-SolrCloud-tp4170624.html
  Sent from the Solr - User mailing list archive at Nabble.com.




-- 
-
Noble Paul


Re: ERROR StreamingSolrServers 4.10.2

2014-11-24 Thread Joseph V J
Hi Team,

Does this mean that the updates on the other server is not getting
forwarded to leader?
Or can this error be safely ignored?

~Regards
Joe


On Mon, Nov 24, 2014 at 12:24 PM, Joseph V J josephveejay...@gmail.com
wrote:

 Hi,

 I have a production Solr Cloud setup which has been migrated from 4.2 to
 4.10.2. Upon then at times I'm getting this ERROR.

 ERROR StreamingSolrServers
 org.apache.solr.common.SolrException: Bad Request

 request:
 http://10.0.0.160:8080/solr/profiles/update?update.distrib=TOLEADERdistrib.from=http%3A%2F%2F10.0.2.160%3A8080%2Fsolr%2Fprofiles%2Fwt=javabinversion=2
 at
 org.apache.solr.client.solrj.impl.ConcurrentUpdateSolrServer$Runner.run(ConcurrentUpdateSolrServer.java:241)
 at
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
 at
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
 at java.lang.Thread.run(Thread.java:745)

 The cloud is of single shard single replica setup with Zookeeper 3.4.6 and
 JAVA version used is
 java version 1.7.0_72
 Java(TM) SE Runtime Environment (build 1.7.0_72-b14)
 Java HotSpot(TM) 64-Bit Server VM (build 24.72-b04, mixed mode)

 It would be great if anyone could throw some light here.

 ~Thanks
 Joe