Re: Error occurred when getting solr-core
Hi all, Cores are not loaded to the CoreContainer. my solr.xml configuration is solr solrcloud str name=host127.0.0.1/str int name=hostPort${jetty.port:8983}/int str name=hostContext${hostContext:solr}/str int name=zkClientTimeout${zkClientTimeout:3}/int str name=shareSchema${shareSchema:false}/str bool name=genericCoreNodeNames${genericCoreNodeNames:true}/bool /solrcloud shardHandlerFactory name=shardHandlerFactory class=HttpShardHandlerFactory int name=socketTimeout${socketTimeout:12}/int int name=connTimeout${connTimeout:15000}/int /shardHandlerFactory /solr Please help. Thanks On Mon, Nov 24, 2014 at 11:35 AM, Danesh Kuruppu dknkuru...@gmail.com wrote: Hi all, I am using solr version 4.7.2 I am getting following when communicating with solr server. org.apache.solr.common.SolrException: No such core: db at org.apache.solr.client.solrj.embedded.EmbeddedSolrServer.request(EmbeddedSolrServer.java:112) at org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:118) at org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:116) at org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:102) I am using EmbeddedSorlServer, It is initialized as follows CoreContainer coreContainer = new CoreContainer(solrHome.getPath()); coreContainer.load(); this.server = new EmbeddedSolrServer(coreContainer, db); in solr-home, there is a directory called db. in core.properties file, I define the name and dataDir. Couldn't find the cause for this error. Please help. Thanks Danesh
Reminder: FOSDEM 2015 - Open Source Search Dev Room
Hi, We host a Dev-Room about Open Source Search on this year's FOSDEM 2015 (https://fosdem.org/2015/), taking place on January 31th and February 1st, 2015, in Brussels, Belgium. There is still one more week to submit your talks, so hurry up and submit your talk early! Here is the full CFP as posted a few weeks ago: Search has evolved to be much more than simply full-text search. We now rely on “search engines” for a wide variety of functionality: search as navigation, search as analytics and backend for data visualization and sometimes, dare we say it, as a data store. The purpose of this dev room is to explore the new world of open source search engines: their enhanced functionality, new use cases, feature and architectural deep dives, and the position of search in relation to the wider set of software tools. We welcome proposals from folks working with or on open source search engines (e.g. Apache Lucene, Apache Solr, Elasticsearch, Seeks, Sphinx, etc.) or technologies that heavily depend upon search (e.g. NoSQL databases, Nutch, Apache Hadoop). We are particularly interested in presentations on search algorithms, machine learning, real-world implementation/deployment stories and explorations of the future of search. Talks should be 30-60 minutes in length, including time for QA. You can submit your talks to us here: https://docs.google.com/forms/d/11yLMj9ZlRD1EMU3Knp5y6eO3H5BRK7V38G0OxSfp84A/viewform Our Call for Papers will close at 23:59 CEST on Monday, December 1, 2014. We cannot guarantee we will have the opportunity to review submissions made after the deadline, so please submit early (and often)! Should you have any questions, you can contact the Dev Room organizers: opensourcesearch-devr...@lists.fosdem.org Cheers, LH on behalf of the Open Source Search Dev Room Program Committee* * Boaz Leskes, Isabel Drost-Fromm, Leslie Hawthorn, Ted Dunning, Torsten Curdt, Uwe Schindler - Uwe Schindler uschind...@apache.org Apache Lucene PMC Member / Committer Bremen, Germany http://lucene.apache.org/
Re: Error occurred when getting solr-core
Hi all, Found the RC for this error. I missed some dependency library for sore-core. Thanks Danesh On Mon, Nov 24, 2014 at 1:43 PM, Danesh Kuruppu dknkuru...@gmail.com wrote: Hi all, Cores are not loaded to the CoreContainer. my solr.xml configuration is solr solrcloud str name=host127.0.0.1/str int name=hostPort${jetty.port:8983}/int str name=hostContext${hostContext:solr}/str int name=zkClientTimeout${zkClientTimeout:3}/int str name=shareSchema${shareSchema:false}/str bool name=genericCoreNodeNames${genericCoreNodeNames:true}/bool /solrcloud shardHandlerFactory name=shardHandlerFactory class=HttpShardHandlerFactory int name=socketTimeout${socketTimeout:12}/int int name=connTimeout${connTimeout:15000}/int /shardHandlerFactory /solr Please help. Thanks On Mon, Nov 24, 2014 at 11:35 AM, Danesh Kuruppu dknkuru...@gmail.com wrote: Hi all, I am using solr version 4.7.2 I am getting following when communicating with solr server. org.apache.solr.common.SolrException: No such core: db at org.apache.solr.client.solrj.embedded.EmbeddedSolrServer.request(EmbeddedSolrServer.java:112) at org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:118) at org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:116) at org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:102) I am using EmbeddedSorlServer, It is initialized as follows CoreContainer coreContainer = new CoreContainer(solrHome.getPath()); coreContainer.load(); this.server = new EmbeddedSolrServer(coreContainer, db); in solr-home, there is a directory called db. in core.properties file, I define the name and dataDir. Couldn't find the cause for this error. Please help. Thanks Danesh
Re: Too much data after closed for HttpChannelOverHttp
It seems to be working now. - I have increased some values in jetty.xml: *Set name=responseBufferSize65536/Set Set name=headerBufferSize32768/Set * *Set name=responseBufferSize32768/Set * - And have left chunk size = 3000. - But I'm not adding all the documents at the same time; instead items are sent one by one to be indexed (it's either that or the jetty configuration altered which is preventing the previous error from being reproduced). Thanks everyone. 2014-11-24 2:36 GMT+01:00 Alexandre Rafalovitch arafa...@gmail.com: Good point on that one Steve. Wireshark is both a hammer and a power drill of network troubleshooting. Takes steady hands to hold it right (it has a bit of a learning curve) but it is a great tool. I swore by it (well Ethereal back then) in my tech support days. So, seconded to try using that if the simple approach fails outright. Regards, Alex. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 23 November 2014 at 20:31, steve sc_shep...@hotmail.com wrote: For what it's worth, depending on the type of PC/MAC you're using, you can use WireShark to look at active http header (sent and received) that are being created for the request. https://www.wireshark.org/ I don't have any financial interest in them, but the stuff works! Steve Date: Sun, 23 Nov 2014 20:47:05 +0100 Subject: Re: Too much data after closed for HttpChannelOverHttp From: h.benoud...@gmail.com To: solr-user@lucene.apache.org Actually I'm using a php client (I think it sends a HTTP request to Solr), but you're right tomorrow once I'll get to the office, I'll set chunk size to a smaller value, and will tell you if that was the reason. Thanks. 2014-11-23 19:35 GMT+01:00 Alexandre Rafalovitch arafa...@gmail.com: Most probably just a request that's too large. Have you tried dropping down to 500 items and seeing what happens? Are you using SolrJ to send content to Solr? Or a direct HTTP request? Regards, Alex. P.s. You may also find it useful to read up on the Solr commit and hard vs. soft commits. Check solrconfig.xml in the example distribution. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 23 November 2014 at 12:31, Hakim Benoudjit h.benoud...@gmail.com wrote: Hi there, I have deployed solr with Jetty, and I'm trying to index a quite large amount of items (300K), retreived from a MySQL database (unfortunately I'm not using DIH; I'm doing it manually, by getting items from MySQL and then index them it in Solr). But, I'm not indexing all of those items at the same time; I'm indexing them by chunks of 3K. So, I get the first 3K, index them, then goes to the next 3K chunk to index it. Here is the error I got in jetty logs, I guess it has nothing to do with Mysql: *Does anyone know the meaning of the error 'badMessage: java.lang.IllegalStateException: too much data after closed for HttpChannelOverHttp@5432494a' ?* Thanks for your help, if anything isnt very precise please tell me to explain it (and sorry for my bad english). -- Cordialement, Best regards, Hakim Benoudjit -- Cordialement, Best regards, Hakim Benoudjit -- Cordialement, Best regards, Hakim Benoudjit
Re: Duplicate facets when the handler configuration specifies facet fields
Good to hear you can reproduce it and thank you for entering that issue into JIRA. FWIW, here is the full list of parameters from echoParams=all from my test: params:{ mlt.minwl:2, mm:5, facet:true, mlt.boost:true, mlt.fl:title,summary,body, facet.mincount:1, mlt.mintf:2, f.typedef.facet.limit:15, mlt.qf:title^5 summary^3 body^1, mlt.maxqt:8, f.subtype.facet.limit:15, mlt.mindf:3, echoParams:all, df:id, mlt.count:3, facet.field:primaryId, facet:true, echoParams:all, indent:true, q:*:*, facet.limit:10, facet.field:primaryId, wt:json, rows:0}}, Also, our testing is for an upgrade from solr 4.4 to solr 4.10.2. I found that in our current production 4.4 instance, the params shows the duplicate (facet.field listed twice), but the actual results only has one set of facets. Guessing there must have been some code on the output side that was removing the duplicates in that version of solr. Thanks, Brian On Nov 24, 2014, at 6:04 AM, Alexandre Rafalovitch arafa...@gmail.commailto:arafa...@gmail.com wrote: I can reproduce it. I added your parameters to the default section of the config and then run the following: curl http://localhost:8983/solr/schemaless/select?q=*:*rows=0wt=jsonindent=truefacet=truefacet.field=primaryId2facet.limit=10echoParams=all; I get: -- params:{ f.typedef.facet.limit:15, facet.field:primaryId2, df:_text, f.subtype.facet.limit:15, echoParams:all, facet.mincount:1, rows:0, facet:true, q:*:*, facet.limit:10, facet.field:primaryId2, indent:true, echoParams:all, rows:0, wt:json, facet:true}}, -- This is against Solr 5 build, but I think that bug is there all the way to Solr 4.1. I think I traced the source of the bug too (parameter and default names are just joined together but with the first (override) value both times as shown above). Usually makes no difference to anything, but it looks like faceting component iterates over the elements, not just gets them, so it gets bitten twice. I've created a JIRA for this issue: https://issues.apache.org/jira/browse/SOLR-6780 Regards, Alex. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 21 November 2014 at 18:29, Alexandre Rafalovitch arafa...@gmail.com wrote: Could you add echoParams=all to the query and see what comes back? Currently, you echo the params you sent, would be good to see what they look like after they combine with defaults. Regards, Alex. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 21 November 2014 18:04, Tom Zimmermann zimm.to...@gmail.com wrote: Brian and I are working together to diagnose this issue so I can chime in quickly here as well. These values are defined as part of the the defaults section of the config.
Highlight returning multiValued field as a single string
I've a multiValued field. Querying on it returns a multivalued result set. But, doing a highlight query returns a single string. I also have hl.preserveMulti=true option. the schema of the field is as follows: field name=some_field type=some_field_type indexed=true stored=true multiValued=true termVectors=true termPositions=true termOffsets=true / fieldType name=some_field_type class=solr.TextField positionIncrementGap=100 analyzer type=index tokenizer class=solr.WhitespaceTokenizerFactory/ filter class=solr.WordDelimiterFilterFactory generateWordParts=1 generateNumberParts=1 catenateWords=1 catenateNumbers=1 catenateAll=0 splitOnCaseChange=1 / filter class=solr.LowerCaseFilterFactory/ /analyzer analyzer type=query tokenizer class=solr.WhitespaceTokenizerFactory/ filter class=solr.SynonymFilterFactory synonyms=synonyms.txt ignoreCase=true expand=true/ filter class=solr.StopFilterFactory ignoreCase=true words=lang/stopwords_en.txt enablePositionIncrements=true / filter class=solr.WordDelimiterFilterFactory generateWordParts=1 generateNumberParts=1 catenateWords=0 catenateNumbers=0 catenateAll=0 splitOnCaseChange=1/ filter class=solr.LowerCaseFilterFactory/ /analyzer /fieldType -- View this message in context: http://lucene.472066.n3.nabble.com/Highlight-returning-multiValued-field-as-a-single-string-tp4170587.html Sent from the Solr - User mailing list archive at Nabble.com.
RE: Lucene ancient greek normalization
If you are using Solr, you can configure your analysis chain to use the ICUFoldingFilterFactory (https://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters#solr.ICUFoldingFilterFactory) and then view the results in the solr admin window. If you are in pure Lucene (circa version 4.8, some mods will be required depending on your version): 1) Extend Analyzer: @Override protected TokenStreamComponents createComponents(String field, Reader reader) { Tokenizer stream = new StandardTokenizer(version, reader); TokenFilter icu = new ICUFoldingFilter(stream); return new TokenStreamComponents(stream, icu); } 2) Then iterate through the tokens: TokenStream stream = analyzer.tokenStream(, new StringReader(text)); stream.reset(); CharTermAttribute cattr = stream.getAttribute(CharTermAttribute.class); while (stream.incrementToken()) { String token = cattr.toString(); ... -Original Message- From: paolo anghileri [mailto:paolo.anghil...@codegeneration.it] Sent: Saturday, November 22, 2014 11:41 AM To: Allison, Timothy B. Subject: Re: Lucene ancient greek normalization Sorry Timothy for the beginner question, how did you manage to run this test? Many thanks Paolo On 21/11/2014 21:14, Allison, Timothy B. wrote: ICU looks promising: Μῆνιν ἄειδε, θεὰ, Πηληϊάδεω Ἀχιλλῆος - 1.μηνιν 2.αειδε 3.θεα 4.πηληιαδεω 5.αχιλληοσ -Original Message- From: Alexandre Rafalovitch [mailto:arafa...@gmail.com] Sent: Friday, November 21, 2014 3:08 PM To: d...@lucene.apache.org Subject: Re: Lucene ancient greek normalization Are you sure that's not something that's already addressed by the ICU Filter? http://www.solr-start.com/javadoc/solr-lucene/org/apache/lucene/analysis/icu/ICUTransformFilterFactory.html If you follow the links to what's possible, the page talks about Greek, though not ancient: http://userguide.icu-project.org/transforms/general#TOC-Greek There was also some discussion on: https://issues.apache.org/jira/browse/LUCENE-1343 Regards, Alex. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 21 November 2014 14:14, paolo anghileri paolo.anghil...@codegeneration.it wrote: For development purposes I need the ability in lucene to normalize ancient greek characters for al the cases of grammatical details such as accents, diacritics and so on. My need is to retrieve ancient greek words with accents and other grammatical details by the input of the string without accents. For example the input of οργανον (organon) should to retrieve also Ὄργανον, I am not a lucene commiter and I a new to this so my question is about the best practice to implement this in Lucene, and possibile submit a commit proposal to Lucene A project management committee. I have made some searches and found this file in Lucene-soir: It contains normalization for some chars. My thought would be to add extra normalization here, including all unicode ancient greek chars with all grammatical details. I already have all the unicode values for that chars so It should not be difficult for me to include them If my understanding is correct, this should add to lucene the features described above. As I am new to this, my needs are: To be sure that this is the correct place in Lucene for doing normalization How to post commit proposal Any help appreciated Kind regards Paolo - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org - To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org
Re: Indexing problems with BBoxField
OK, David Smiley now already created an issue for this: https://issues.apache.org/jira/browse/SOLR-6781 So it really is a bug. Furthermore, I also had a lot of problems trying to search on the field after finally getting it indexed. I summarized those here: https://issues.apache.org/jira/browse/SOLR-6784 On 2014-11-23 14:34, re...@gmx.net wrote: Thanks a lot for your reply! I had »docValues=true« in there before, but then thought I'd try out removing it to see if that helped. It didn't, and I forgot to re-add it before copying it into the mail. So, unfortunately, that's not it. However, the other one seems to bring us a step closer to the solution: After adding field name=bboxs_field_location_area type=bbox indexed=true stored=false multiValued=false/ (even without removing the dynamic fields), this works indeed just fine! So, the question is what causes this, and it seems more and more like a bug instead of a user error. But I'll wait for a bit more feedback before filing a Jira. On 2014-11-23 14:10, Jack Krupansky wrote: A difference I see in your snippet from the example is that you don't have docValues=true on the coordinate field type. You wrote: fieldType name=_bbox_coord class=solr.TrieDoubleField precisionStep=8 stored=false / But the example is: fieldType name=_bbox_coord class=solr.TrieDoubleField precisionStep=8 docValues=true stored=false/ Also, maybe try a static field rather than dynamic field, although the latter should work anyway. Please file a Jira to request that Solr give a user-sensible error, not a Lucene-level error. I mean, the Solr user has no ability to directly invoke the createFields method. And now... let's see what David Smiley has to say about all of this! -- Jack Krupansky -Original Message- From: Thomas Seidl Sent: Sunday, November 23, 2014 6:33 AM To: solr-user@lucene.apache.org Subject: Indexing problems with BBoxField Hi all, I just downloaded Solr 4.10.2 and wanted to try out the new BBoxField type, but couldn't get it to work. The error (with status 400) I get is: ERROR: [doc=foo] Error adding field 'bboxs_field_location_area'='ENVELOPE(25.89, 41.13, 47.07, 35.31)' msg=java.lang.IllegalStateException: instead call createFields() because isPolyField() is true Which, of course, is rather unhelpful for a user. The relevant portions of my schema.xml look like this (largely copied from [1]: fieldType name=bbox class=solr.BBoxField geo=true units=degrees numberType=_bbox_coord / fieldType name=_bbox_coord class=solr.TrieDoubleField precisionStep=8 stored=false / dynamicField name=bboxs_* type=bbox indexed=true stored=false multiValued=false/ [1] https://cwiki.apache.org/confluence/display/solr/Spatial+Search And the request I send is this: add doc field name=idfoo/field field name=bboxs_field_location_areaENVELOPE(25.89, 41.13, 47.07, 35.31)/field /doc /add Does anyone have any idea what could be going wrong here? Thanks a lot in advance, Thomas
Re: Indexing problems with BBoxField
Thomas, Thank you for communicating on the list about your experience and raising the JIRA issue. I meant to respond last night but lost the chance. (and Jack, thanks for helping Thomas out). I’ll follow-up to SOLR-6784. SOLR-6781 now has a bug-fix patch. I’ll apply it later today. ~ David Smiley Freelance Apache Lucene/Solr Search Consultant/Developer http://www.linkedin.com/in/davidwsmiley On Mon, Nov 24, 2014 at 8:43 AM, re...@gmx.net wrote: OK, David Smiley now already created an issue for this: https://issues.apache.org/jira/browse/SOLR-6781 So it really is a bug. Furthermore, I also had a lot of problems trying to search on the field after finally getting it indexed. I summarized those here: https://issues.apache.org/jira/browse/SOLR-6784 On 2014-11-23 14:34, re...@gmx.net wrote: Thanks a lot for your reply! I had »docValues=true« in there before, but then thought I'd try out removing it to see if that helped. It didn't, and I forgot to re-add it before copying it into the mail. So, unfortunately, that's not it. However, the other one seems to bring us a step closer to the solution: After adding field name=bboxs_field_location_area type=bbox indexed=true stored=false multiValued=false/ (even without removing the dynamic fields), this works indeed just fine! So, the question is what causes this, and it seems more and more like a bug instead of a user error. But I'll wait for a bit more feedback before filing a Jira. On 2014-11-23 14:10, Jack Krupansky wrote: A difference I see in your snippet from the example is that you don't have docValues=true on the coordinate field type. You wrote: fieldType name=_bbox_coord class=solr.TrieDoubleField precisionStep=8 stored=false / But the example is: fieldType name=_bbox_coord class=solr.TrieDoubleField precisionStep=8 docValues=true stored=false/ Also, maybe try a static field rather than dynamic field, although the latter should work anyway. Please file a Jira to request that Solr give a user-sensible error, not a Lucene-level error. I mean, the Solr user has no ability to directly invoke the createFields method. And now... let's see what David Smiley has to say about all of this! -- Jack Krupansky -Original Message- From: Thomas Seidl Sent: Sunday, November 23, 2014 6:33 AM To: solr-user@lucene.apache.org Subject: Indexing problems with BBoxField Hi all, I just downloaded Solr 4.10.2 and wanted to try out the new BBoxField type, but couldn't get it to work. The error (with status 400) I get is: ERROR: [doc=foo] Error adding field 'bboxs_field_location_area'='ENVELOPE(25.89, 41.13, 47.07, 35.31)' msg=java.lang.IllegalStateException: instead call createFields() because isPolyField() is true Which, of course, is rather unhelpful for a user. The relevant portions of my schema.xml look like this (largely copied from [1]: fieldType name=bbox class=solr.BBoxField geo=true units=degrees numberType=_bbox_coord / fieldType name=_bbox_coord class=solr.TrieDoubleField precisionStep=8 stored=false / dynamicField name=bboxs_* type=bbox indexed=true stored=false multiValued=false/ [1] https://cwiki.apache.org/confluence/display/solr/Spatial+Search And the request I send is this: add doc field name=idfoo/field field name=bboxs_field_location_areaENVELOPE(25.89, 41.13, 47.07, 35.31)/field /doc /add Does anyone have any idea what could be going wrong here? Thanks a lot in advance, Thomas
use {!surround} with more conditions
Can't make surround work with other conditions... for example: {!surround}SomeField:30w(Text1*,Text2*) -SomeField:Text3* SomeField:* -{!surround}SomeField:30w(Text1*,Text2*) Is that doable?
Missing value with Date Range
Hi All, I'm trying to get missing count with Date Range adding facet.missing=true as parameter but this not work (which work with facet). Do you have idea how to get missing count with date range or numeric range. Regards,Nabil.
Backuping SolrCloud
Hi, I'm looking for a built-in SolrCloud backup mechanism. I want to backup my Index (scheduled / manual backups) while Indexing and searching. What is the proper way to perform this backup-restore task? Thanks. -- View this message in context: http://lucene.472066.n3.nabble.com/Backuping-SolrCloud-tp4170624.html Sent from the Solr - User mailing list archive at Nabble.com.
how solr4.0 and zookeeper run on weblogic
Hi , Is there a way to deploy Zookeeper on weblogic cluster. Thanks and Regards Kunal Sharma 07917836122
CoreContainer : create new cores reusing/sharing solrconfig.xml and schema.xml
Given the following directory/file structure .../solrhome .../solrhome/solr.xml .../solrhome/config/ .../solrhome/config/schema.xml .../solrhome/config/solrconfig.xml .../solrhome/cores How can I programmatically (CoreContainer) create a new core below .../solrhome/cores that makes use of solrconfig.xml and schema.xml? The new core should use the config files in .../solrhome/config/ Thx Clemens
Re: CoreContainer : create new cores reusing/sharing solrconfig.xml and schema.xml
Have you tried Configsets: https://cwiki.apache.org/confluence/display/solr/Config+Sets ? Regards, Alex. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 24 November 2014 at 10:14, Clemens Wyss DEV clemens...@mysign.ch wrote: Given the following directory/file structure .../solrhome .../solrhome/solr.xml .../solrhome/config/ .../solrhome/config/schema.xml .../solrhome/config/solrconfig.xml .../solrhome/cores How can I programmatically (CoreContainer) create a new core below .../solrhome/cores that makes use of solrconfig.xml and schema.xml? The new core should use the config files in .../solrhome/config/ Thx Clemens
AW: CoreContainer : create new cores reusing/sharing solrconfig.xml and schema.xml
Thanks fort he quick reply. But how do I hand in the configSet to use. I guess through the CoreDescriptor? -Ursprüngliche Nachricht- Von: Alexandre Rafalovitch [mailto:arafa...@gmail.com] Gesendet: Montag, 24. November 2014 16:27 An: solr-user Betreff: Re: CoreContainer : create new cores reusing/sharing solrconfig.xml and schema.xml Have you tried Configsets: https://cwiki.apache.org/confluence/display/solr/Config+Sets ? Regards, Alex. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 24 November 2014 at 10:14, Clemens Wyss DEV clemens...@mysign.ch wrote: Given the following directory/file structure .../solrhome .../solrhome/solr.xml .../solrhome/config/ .../solrhome/config/schema.xml .../solrhome/config/solrconfig.xml .../solrhome/cores How can I programmatically (CoreContainer) create a new core below .../solrhome/cores that makes use of solrconfig.xml and schema.xml? The new core should use the config files in .../solrhome/config/ Thx Clemens
Re: Backuping SolrCloud
I was able to get very good backup procedure by having cron job perform compact on each shard and then copy out the physical shard (the full lucene index directory to a backup server) Updates would need to be stopped during this period. And restore would be as simple as copying in the backed up shard and restarting solr On Nov 24, 2014, at 9:48 AM, elmerfudd na...@012.net.il wrote: Hi, I'm looking for a built-in SolrCloud backup mechanism. I want to backup my Index (scheduled / manual backups) while Indexing and searching. What is the proper way to perform this backup-restore task? Thanks. -- View this message in context: http://lucene.472066.n3.nabble.com/Backuping-SolrCloud-tp4170624.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: CoreContainer : create new cores reusing/sharing solrconfig.xml and schema.xml
I haven't done this myself, but it is defined as a constant in the CoreDescriptor: http://www.solr-start.com/javadoc/solr-lucene/org/apache/solr/core/CoreDescriptor.html#CORE_CONFIGSET You could also look at the tests and see whether it is used. Regards, Alex. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 24 November 2014 at 10:43, Clemens Wyss DEV clemens...@mysign.ch wrote: Thanks fort he quick reply. But how do I hand in the configSet to use. I guess through the CoreDescriptor? -Ursprüngliche Nachricht- Von: Alexandre Rafalovitch [mailto:arafa...@gmail.com] Gesendet: Montag, 24. November 2014 16:27 An: solr-user Betreff: Re: CoreContainer : create new cores reusing/sharing solrconfig.xml and schema.xml Have you tried Configsets: https://cwiki.apache.org/confluence/display/solr/Config+Sets ? Regards, Alex. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 24 November 2014 at 10:14, Clemens Wyss DEV clemens...@mysign.ch wrote: Given the following directory/file structure .../solrhome .../solrhome/solr.xml .../solrhome/config/ .../solrhome/config/schema.xml .../solrhome/config/solrconfig.xml .../solrhome/cores How can I programmatically (CoreContainer) create a new core below .../solrhome/cores that makes use of solrconfig.xml and schema.xml? The new core should use the config files in .../solrhome/config/ Thx Clemens
AW: CoreContainer : create new cores reusing/sharing solrconfig.xml and schema.xml
Looks like 4.7.2 does not have this const. Another reason to switch to 4.10 (and java 8 ;) ) -Ursprüngliche Nachricht- Von: Alexandre Rafalovitch [mailto:arafa...@gmail.com] Gesendet: Montag, 24. November 2014 16:49 An: solr-user Betreff: Re: CoreContainer : create new cores reusing/sharing solrconfig.xml and schema.xml I haven't done this myself, but it is defined as a constant in the CoreDescriptor: http://www.solr-start.com/javadoc/solr-lucene/org/apache/solr/core/CoreDescriptor.html#CORE_CONFIGSET You could also look at the tests and see whether it is used. Regards, Alex. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 24 November 2014 at 10:43, Clemens Wyss DEV clemens...@mysign.ch wrote: Thanks fort he quick reply. But how do I hand in the configSet to use. I guess through the CoreDescriptor? -Ursprüngliche Nachricht- Von: Alexandre Rafalovitch [mailto:arafa...@gmail.com] Gesendet: Montag, 24. November 2014 16:27 An: solr-user Betreff: Re: CoreContainer : create new cores reusing/sharing solrconfig.xml and schema.xml Have you tried Configsets: https://cwiki.apache.org/confluence/display/solr/Config+Sets ? Regards, Alex. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 24 November 2014 at 10:14, Clemens Wyss DEV clemens...@mysign.ch wrote: Given the following directory/file structure .../solrhome .../solrhome/solr.xml .../solrhome/config/ .../solrhome/config/schema.xml .../solrhome/config/solrconfig.xml .../solrhome/cores How can I programmatically (CoreContainer) create a new core below .../solrhome/cores that makes use of solrconfig.xml and schema.xml? The new core should use the config files in .../solrhome/config/ Thx Clemens
Re: Backuping SolrCloud
There are some JIRAs out there for better backup/restore for SolrCloud situations, but nothing has been committed yet that I know of. Best, Erick On Mon, Nov 24, 2014 at 7:45 AM, Vivek Pathak vpat...@orgmeta.com wrote: I was able to get very good backup procedure by having cron job perform compact on each shard and then copy out the physical shard (the full lucene index directory to a backup server) Updates would need to be stopped during this period. And restore would be as simple as copying in the backed up shard and restarting solr On Nov 24, 2014, at 9:48 AM, elmerfudd na...@012.net.il wrote: Hi, I'm looking for a built-in SolrCloud backup mechanism. I want to backup my Index (scheduled / manual backups) while Indexing and searching. What is the proper way to perform this backup-restore task? Thanks. -- View this message in context: http://lucene.472066.n3.nabble.com/Backuping-SolrCloud-tp4170624.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Backuping SolrCloud
I have a writeup of how to perform safe backups here: https://gist.github.com/ralph-tice/887414a7f8082a0cb828 There are some tickets around this work to further the ease of backups, especially https://issues.apache.org/jira/browse/SOLR-5750 On Mon, Nov 24, 2014 at 9:45 AM, Vivek Pathak vpat...@orgmeta.com wrote: I was able to get very good backup procedure by having cron job perform compact on each shard and then copy out the physical shard (the full lucene index directory to a backup server) Updates would need to be stopped during this period. And restore would be as simple as copying in the backed up shard and restarting solr On Nov 24, 2014, at 9:48 AM, elmerfudd na...@012.net.il wrote: Hi, I'm looking for a built-in SolrCloud backup mechanism. I want to backup my Index (scheduled / manual backups) while Indexing and searching. What is the proper way to perform this backup-restore task? Thanks. -- View this message in context: http://lucene.472066.n3.nabble.com/Backuping-SolrCloud-tp4170624.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: CoreContainer : create new cores reusing/sharing solrconfig.xml and schema.xml
4.10 does NOT require java 8, it requires Java 1.7. In fact, 5.x is on Java 1.7 too. Trunk, (the future Solr/Lucene 6.0) is on Java 8 though (just recently moved). But if you wan to go to Java 8 anyway and need an excuse pretend I didn't write this ;). Best, Erick On Mon, Nov 24, 2014 at 7:58 AM, Clemens Wyss DEV clemens...@mysign.ch wrote: Looks like 4.7.2 does not have this const. Another reason to switch to 4.10 (and java 8 ;) ) -Ursprüngliche Nachricht- Von: Alexandre Rafalovitch [mailto:arafa...@gmail.com] Gesendet: Montag, 24. November 2014 16:49 An: solr-user Betreff: Re: CoreContainer : create new cores reusing/sharing solrconfig.xml and schema.xml I haven't done this myself, but it is defined as a constant in the CoreDescriptor: http://www.solr-start.com/javadoc/solr-lucene/org/apache/solr/core/CoreDescriptor.html#CORE_CONFIGSET You could also look at the tests and see whether it is used. Regards, Alex. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 24 November 2014 at 10:43, Clemens Wyss DEV clemens...@mysign.ch wrote: Thanks fort he quick reply. But how do I hand in the configSet to use. I guess through the CoreDescriptor? -Ursprüngliche Nachricht- Von: Alexandre Rafalovitch [mailto:arafa...@gmail.com] Gesendet: Montag, 24. November 2014 16:27 An: solr-user Betreff: Re: CoreContainer : create new cores reusing/sharing solrconfig.xml and schema.xml Have you tried Configsets: https://cwiki.apache.org/confluence/display/solr/Config+Sets ? Regards, Alex. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 24 November 2014 at 10:14, Clemens Wyss DEV clemens...@mysign.ch wrote: Given the following directory/file structure .../solrhome .../solrhome/solr.xml .../solrhome/config/ .../solrhome/config/schema.xml .../solrhome/config/solrconfig.xml .../solrhome/cores How can I programmatically (CoreContainer) create a new core below .../solrhome/cores that makes use of solrconfig.xml and schema.xml? The new core should use the config files in .../solrhome/config/ Thx Clemens
Re: use {!surround} with more conditions
Hi, Not sure about SomeField:* but others should work. You can move clauses that do not require surround to filter queries. e.g. fq=-SomeField:Text3* Ahmet On Monday, November 24, 2014 4:33 PM, Bet Hel beteli8...@gmail.com wrote: Can't make surround work with other conditions... for example: {!surround}SomeField:30w(Text1*,Text2*) -SomeField:Text3* SomeField:* -{!surround}SomeField:30w(Text1*,Text2*) Is that doable?
Re: how solr4.0 and zookeeper run on weblogic
I hope this post helps you http://lucene.472066.n3.nabble.com/how-solr4-0-and-zookeeper-run-on-weblogic-td4013882.html -- View this message in context: http://lucene.472066.n3.nabble.com/how-solr4-0-and-zookeeper-run-on-weblogic-tp4170625p4170679.html Sent from the Solr - User mailing list archive at Nabble.com.
Documents to query
Hi Is there a way (in Lucene / SOLR) to go from a vector of documents (IDs, and optionally their scores) to a query? i.e. the reverse process of query - document? thank you Peyman
Re: Documents to query
Very unlikely. What's the business case? Regards, Alex. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 24 November 2014 at 13:23, Robust Links pey...@robustlinks.com wrote: Hi Is there a way (in Lucene / SOLR) to go from a vector of documents (IDs, and optionally their scores) to a query? i.e. the reverse process of query - document? thank you Peyman
Re: Documents to query
This feature is called “more like this”. I think it only works for a single document, but it probably could be extended. wunder Walter Underwood wun...@wunderwood.org http://observer.wunderwood.org/ On Nov 24, 2014, at 10:26 AM, Alexandre Rafalovitch arafa...@gmail.com wrote: Very unlikely. What's the business case? Regards, Alex. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 24 November 2014 at 13:23, Robust Links pey...@robustlinks.com wrote: Hi Is there a way (in Lucene / SOLR) to go from a vector of documents (IDs, and optionally their scores) to a query? i.e. the reverse process of query - document? thank you Peyman
matching shingles issue
I have Solr 4.2.1 I am using the following analyser: fieldType name=text_shingle class=solr.TextField positionIncrementGap=100 analyzer type=index tokenizer class=solr.WhitespaceTokenizerFactory/ filter class=solr.ShingleFilterFactory minShingleSize=2 maxShingleSize=5 outputUnigrams=true outputUnigramsIfNoShingles=false tokenSeparator= / /analyzer analyzer type=query tokenizer class=solr.WhitespaceTokenizerFactory/ filter class=solr.ShingleFilterFactory minShingleSize=2 maxShingleSize=5 outputUnigrams=false outputUnigramsIfNoShingles=true tokenSeparator= / /analyzer /fieldType for Query: description_shingle:Highest quality I am getting Result: arr name=description_shingle strHighest standards of quality installations!/str /arr So the result does not have shingle Highest quality Instead it has Highest standards of quality The question is why I am getting this match -- View this message in context: http://lucene.472066.n3.nabble.com/matching-shingles-issue-tp4170685.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Documents to query
@ Alexandre.. the business case is the following. I am using lucene/solr to compose vectors. I want to apply different functions (addition, multipication, tensor products,etc) to these vectors. The end result of that is a new vector. I would like to then go back from that vector back to term that might have generated that vector..i.e. the following pipeline 1) term i - query i - N docs_of_i 2) term j - query j - N docs_of_j 3) F(docs_of_i, docs_of_j) - N docs 4) N docs - term_k 1 2 are usual query path. steps 3 and 4 are what i am trying to do @walter. if I am correct, MLT will generate a term for *EACH* element of the vector (N docs above).. but I am interested in a single term On Mon, Nov 24, 2014 at 1:31 PM, Walter Underwood wun...@wunderwood.org wrote: This feature is called “more like this”. I think it only works for a single document, but it probably could be extended. wunder Walter Underwood wun...@wunderwood.org http://observer.wunderwood.org/ On Nov 24, 2014, at 10:26 AM, Alexandre Rafalovitch arafa...@gmail.com wrote: Very unlikely. What's the business case? Regards, Alex. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 24 November 2014 at 13:23, Robust Links pey...@robustlinks.com wrote: Hi Is there a way (in Lucene / SOLR) to go from a vector of documents (IDs, and optionally their scores) to a query? i.e. the reverse process of query - document? thank you Peyman
Re: Indexing with SolrJ fails on windows
The problem seems to occur at the apache redirect - I found if I bypass apache by using my VM IP address directly as the Solr URL, then the error does not occur even from windows. From some searching it seems like Apache does not allow carriage returns in its request headers - so my guess is SolrJ is adding a carriage return in the request when run on Windows, but not on linux, so when receiving the request from Windows apache is spitting back an error. Is there any way to disable this behavior with SolrJ / have it generate consistent requests regardless of platform? -- View this message in context: http://lucene.472066.n3.nabble.com/Indexing-with-SolrJ-fails-on-windows-tp4170687p4170690.html Sent from the Solr - User mailing list archive at Nabble.com.
Indexing with SolrJ fails on windows
I am using solrj to index to Solr through a Java application - I've tried this both with Solr 4.8.1 and Solr 4.10.2, indexing to Solr 4.10.0. I've found I cannot index too large content (a field with 400 words) or more than 1 document at once to Solr instances from windows. The exact same indexing code works from linux, unchanged. I've found it does not work on Windows if: 1) I try to add more than 1 document at a time 2) I try to add a document with a long field value (400 words). However in both cases it works fine if run from linux, or on windows if I only add one document without very long values for any field. The exception I get is the following: org.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException: Invalid chunk header at org.apache.solr.client.solrj.impl.HttpSolrServer.executeMethod(HttpSolrServer.java:552) ~[solr-solrj-4.10.2.jar:4.10.2 1634293 - mike - 2014-10-26 05:56:22] at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:210) ~[solr-solrj-4.10.2.jar:4.10.2 1634293 - mike - 2014-10-26 05:56:22] at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:206) ~[solr-solrj-4.10.2.jar:4.10.2 1634293 - mike - 2014-10-26 05:56:22] at org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:124) ~[solr-solrj-4.10.2.jar:4.10.2 1634293 - mike - 2014-10-26 05:56:22] at org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:68) ~[solr-solrj-4.10.2.jar:4.10.2 1634293 - mike - 2014-10-26 05:56:22] at org.apache.solr.client.solrj.SolrServer.add(SolrServer.java:54) ~[solr-solrj-4.10.2.jar:4.10.2 1634293 - mike - 2014-10-26 05:56:22] My test code to reproduce this is the following: @Test public void testWriteBigOther() throws Exception { SolrServer solrServer = new HttpSolrServer(http://my-vm:8080/solr/test_copy;); SolrInputDocument doc = new SolrInputDocument(); doc.addField(asset_id,test_write_big); doc.addField(title, test write big title); doc.addField(secondary_header, StringUtils.repeat(396, ,400)); ListSolrInputDocument inputDocs = new ArrayListSolrInputDocument(); inputDocs.add(doc); solrServer.add(inputDocs); solrServer.commit(); solrServer.shutdown(); } (it uses org.apache.commons.lang3.StringUtils repeat() method to generate the large field value). It seems like there must be a bug in SolrJ - i.e., I guess when it is building the request it does something differently in windows vs. linux - like maybe adds a carriage return on windows? Does anyone know how to fix this, or what else I could do to diagnose it? -- View this message in context: http://lucene.472066.n3.nabble.com/Indexing-with-SolrJ-fails-on-windows-tp4170687.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Documents to query
Is this the Paris - France + Italy = Rome thing? I think I've seen a couple of discussions about that. When you say a term are you actually looking for one keyword? Because that's very different from a query that can be infinitely complex. Regards, Alex. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 24 November 2014 at 13:44, Robust Links pey...@robustlinks.com wrote: @ Alexandre.. the business case is the following. I am using lucene/solr to compose vectors. I want to apply different functions (addition, multipication, tensor products,etc) to these vectors. The end result of that is a new vector. I would like to then go back from that vector back to term that might have generated that vector..i.e. the following pipeline 1) term i - query i - N docs_of_i 2) term j - query j - N docs_of_j 3) F(docs_of_i, docs_of_j) - N docs 4) N docs - term_k 1 2 are usual query path. steps 3 and 4 are what i am trying to do @walter. if I am correct, MLT will generate a term for *EACH* element of the vector (N docs above).. but I am interested in a single term On Mon, Nov 24, 2014 at 1:31 PM, Walter Underwood wun...@wunderwood.org wrote: This feature is called “more like this”. I think it only works for a single document, but it probably could be extended. wunder Walter Underwood wun...@wunderwood.org http://observer.wunderwood.org/ On Nov 24, 2014, at 10:26 AM, Alexandre Rafalovitch arafa...@gmail.com wrote: Very unlikely. What's the business case? Regards, Alex. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 24 November 2014 at 13:23, Robust Links pey...@robustlinks.com wrote: Hi Is there a way (in Lucene / SOLR) to go from a vector of documents (IDs, and optionally their scores) to a query? i.e. the reverse process of query - document? thank you Peyman
Re: matching shingles issue
maybe try description_shingle:(Highest quality) On 11/24/14 1:46 PM, vit wrote: I have Solr 4.2.1 I am using the following analyser: fieldType name=text_shingle class=solr.TextField positionIncrementGap=100 analyzer type=index tokenizer class=solr.WhitespaceTokenizerFactory/ filter class=solr.ShingleFilterFactory minShingleSize=2 maxShingleSize=5 outputUnigrams=true outputUnigramsIfNoShingles=false tokenSeparator= / /analyzer analyzer type=query tokenizer class=solr.WhitespaceTokenizerFactory/ filter class=solr.ShingleFilterFactory minShingleSize=2 maxShingleSize=5 outputUnigrams=false outputUnigramsIfNoShingles=true tokenSeparator= / /analyzer /fieldType for Query: description_shingle:Highest quality I am getting Result: arr name=description_shingle strHighest standards of quality installations!/str /arr So the result does not have shingle Highest quality Instead it has Highest standards of quality The question is why I am getting this match -- View this message in context: http://lucene.472066.n3.nabble.com/matching-shingles-issue-tp4170685.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Indexing with SolrJ fails on windows
Try running the client with -Dline.separator='\n' to force the line separator. https://docs.oracle.com/javase/tutorial/essential/environment/sysprop.html However, if it that does work, it's probably a bug. Regards, Alex. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 24 November 2014 at 14:00, brian4 bqu...@gmail.com wrote: The problem seems to occur at the apache redirect - I found if I bypass apache by using my VM IP address directly as the Solr URL, then the error does not occur even from windows. From some searching it seems like Apache does not allow carriage returns in its request headers - so my guess is SolrJ is adding a carriage return in the request when run on Windows, but not on linux, so when receiving the request from Windows apache is spitting back an error. Is there any way to disable this behavior with SolrJ / have it generate consistent requests regardless of platform? -- View this message in context: http://lucene.472066.n3.nabble.com/Indexing-with-SolrJ-fails-on-windows-tp4170687p4170690.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Documents to query
i term is the title of the document indexed. So its not an unbounded query term. On Mon, Nov 24, 2014 at 2:02 PM, Alexandre Rafalovitch arafa...@gmail.com wrote: Is this the Paris - France + Italy = Rome thing? I think I've seen a couple of discussions about that. When you say a term are you actually looking for one keyword? Because that's very different from a query that can be infinitely complex. Regards, Alex. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 24 November 2014 at 13:44, Robust Links pey...@robustlinks.com wrote: @ Alexandre.. the business case is the following. I am using lucene/solr to compose vectors. I want to apply different functions (addition, multipication, tensor products,etc) to these vectors. The end result of that is a new vector. I would like to then go back from that vector back to term that might have generated that vector..i.e. the following pipeline 1) term i - query i - N docs_of_i 2) term j - query j - N docs_of_j 3) F(docs_of_i, docs_of_j) - N docs 4) N docs - term_k 1 2 are usual query path. steps 3 and 4 are what i am trying to do @walter. if I am correct, MLT will generate a term for *EACH* element of the vector (N docs above).. but I am interested in a single term On Mon, Nov 24, 2014 at 1:31 PM, Walter Underwood wun...@wunderwood.org wrote: This feature is called “more like this”. I think it only works for a single document, but it probably could be extended. wunder Walter Underwood wun...@wunderwood.org http://observer.wunderwood.org/ On Nov 24, 2014, at 10:26 AM, Alexandre Rafalovitch arafa...@gmail.com wrote: Very unlikely. What's the business case? Regards, Alex. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 24 November 2014 at 13:23, Robust Links pey...@robustlinks.com wrote: Hi Is there a way (in Lucene / SOLR) to go from a vector of documents (IDs, and optionally their scores) to a query? i.e. the reverse process of query - document? thank you Peyman
Help on Notification for SOLR
Hello, I am working on implementing a notification system when there is any change in the solr index (Addition or removal of documents from Index). Can someone guide me on the starting point. I see some articles using DIH? Is there any way using without DIH. Thanks Ravi
Re: Documents to query
Doesn't it sound like faceting? On Mon, Nov 24, 2014 at 9:44 PM, Robust Links pey...@robustlinks.com wrote: @ Alexandre.. the business case is the following. I am using lucene/solr to compose vectors. I want to apply different functions (addition, multipication, tensor products,etc) to these vectors. The end result of that is a new vector. I would like to then go back from that vector back to term that might have generated that vector..i.e. the following pipeline 1) term i - query i - N docs_of_i 2) term j - query j - N docs_of_j 3) F(docs_of_i, docs_of_j) - N docs 4) N docs - term_k 1 2 are usual query path. steps 3 and 4 are what i am trying to do @walter. if I am correct, MLT will generate a term for *EACH* element of the vector (N docs above).. but I am interested in a single term On Mon, Nov 24, 2014 at 1:31 PM, Walter Underwood wun...@wunderwood.org wrote: This feature is called “more like this”. I think it only works for a single document, but it probably could be extended. wunder Walter Underwood wun...@wunderwood.org http://observer.wunderwood.org/ On Nov 24, 2014, at 10:26 AM, Alexandre Rafalovitch arafa...@gmail.com wrote: Very unlikely. What's the business case? Regards, Alex. Personal: http://www.outerthoughts.com/ and @arafalov Solr resources and newsletter: http://www.solr-start.com/ and @solrstart Solr popularizers community: https://www.linkedin.com/groups?gid=6713853 On 24 November 2014 at 13:23, Robust Links pey...@robustlinks.com wrote: Hi Is there a way (in Lucene / SOLR) to go from a vector of documents (IDs, and optionally their scores) to a query? i.e. the reverse process of query - document? thank you Peyman -- Sincerely yours Mikhail Khludnev Principal Engineer, Grid Dynamics http://www.griddynamics.com mkhlud...@griddynamics.com
RE: Help on Notification for SOLR
Hi, You might want to take a look at the UpdateRequestProcessor and UpdateRequestProcessorChain classes: https://wiki.apache.org/solr/UpdateRequestProcessor and http://lucene.apache.org/solr/4_10_2/solr-core/org/apache/solr/update/processor/UpdateRequestProcessorChain.html You can hook a chain to your update request handler and use a ScriptUpdateProcessor to implement your notification system. https://wiki.apache.org/solr/ScriptUpdateProcessor Hope this helps, Gonzalo -Original Message- From: EXTERNAL Taminidi Ravi (ETI, Automotive-Service-Solutions) [mailto:external.ravi.tamin...@us.bosch.com] Sent: Monday, November 24, 2014 2:35 PM To: solr-user@lucene.apache.org Subject: Help on Notification for SOLR Hello, I am working on implementing a notification system when there is any change in the solr index (Addition or removal of documents from Index). Can someone guide me on the starting point. I see some articles using DIH? Is there any way using without DIH. Thanks Ravi
Re: Help on Notification for SOLR
There are commit notifications scripts in solrconfig.xml, there is JMX, there is whatever the admin UI uses for freeze and show change screen. And, as mentioned already, URP chains. Regards, Alex On 24/11/2014 3:36 pm, EXTERNAL Taminidi Ravi (ETI, Automotive-Service-Solutions) external.ravi.tamin...@us.bosch.com wrote: Hello, I am working on implementing a notification system when there is any change in the solr index (Addition or removal of documents from Index). Can someone guide me on the starting point. I see some articles using DIH? Is there any way using without DIH. Thanks Ravi
Re: Help on Notification for SOLR
of course there is another approach which is using Solr Clients. So, you can use solr clients for specific language like PHP, JAVA, ... For Example : JAVA Solr Client : SolrJ is an API that makes it easy for Java applications to talk to Solr. SolrJ hides a lot of the details of connecting to Solr and allows your application to interact with Solr with simple high-level methods. String urlString = http://localhost:8983/solr;; SolrServer solr = new HttpSolrServer(urlString); SolrInputDocument document = new SolrInputDocument(); document.addField(id, 552199); document.addField(name, Gouda cheese wheel); document.addField(price, 49.99); UpdateResponse response = solr.add(document); // Remember to commit your changes! solr.commit(); If you want another language, you can tell me and I will help you . EXTERNAL Taminidi Ravi (ETI, Automotive-Service-Solutions) wrote Hello, I am working on implementing a notification system when there is any change in the solr index (Addition or removal of documents from Index). Can someone guide me on the starting point. I see some articles using DIH? Is there any way using without DIH. Thanks Ravi -- View this message in context: http://lucene.472066.n3.nabble.com/Help-on-Notification-for-SOLR-tp4170708p4170716.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Using stored value of a field to build suggester index
Thanks Erick, This makes things clearer. Thanks, Faisal On Sun, Nov 23, 2014 at 2:17 PM, Erick Erickson erickerick...@gmail.com wrote: You can't build the suggester from the stored values, it's constructed from indexed terms only. You probably want to create a copyField to a less-analyzed (indexed) field and suggest from _that_. You'll probably want to do things like remove punctuation, perhaps lowercase and the like but not stem etc. Best, Erick On Sun, Nov 23, 2014 at 12:25 PM, Faisal Mansoor faisal.mans...@gmail.com wrote: Hi, I am trying to build a suggester for a field which is both index and stored. The field is whitespace tokenized, lowercased, stemmed etc while indexing. It looks like that the indexed terms are used as a source for building the suggester index. Which is what the following line in the suggester documentation also mentions. https://wiki.apache.org/solr/Suggester - field - if sourceLocation is empty then terms from this field in the index will be used when building the trie. I want to display the suggested value in UI, is it possible to use the stored value of the field rather than the indexed terms to build the index. Here are the relevant definitions from solrconfig.xml and schema.xml. Thanks. Faisal solrconfig.xml searchComponent class=solr.SpellCheckComponent name=infix_suggest_analyzing lst name=spellchecker str name=nameinfix_suggest_analyzing/str str name=classnameorg.apache.solr.spelling.suggest.Suggester/str str name=lookupImplorg.apache.solr.spelling.suggest.fst.AnalyzingInfixLookupFactory/str str name=buildOnCommitfalse/str !-- Suggester properties -- str name=suggestAnalyzerFieldTypeautosuggest_fieldType/str str name=dictionaryImplorg.apache.solr.spelling.suggest.HighFrequencyDictionaryFactory/str str name=fieldDisplayName/str /lst !-- specify a fieldtype using keywordtokenizer + lowercase + cleanup -- str name=queryAnalyzerFieldTypephrase_suggest/str /searchComponent requestHandler name=/suggest class=org.apache.solr.handler.component.SearchHandler lst name=defaults str name=echoParamsexplicit/str str name=spellchecktrue/str str name=spellcheck.dictionaryinfix_suggest_analyzing/str str name=spellcheck.onlyMorePopulartrue/str str name=spellcheck.count200/str str name=spellcheck.collatetrue/str str name=spellcheck.maxCollations10/str /lst arr name=components strinfix_suggest_analyzing/str /arr /requestHandler schema.xml fieldType name=autosuggest_fieldType class=solr.TextField positionIncrementGap=100 analyzer tokenizer class=solr.StandardTokenizerFactory/ filter class=solr.LowerCaseFilterFactory/ filter class=solr.ASCIIFoldingFilterFactory/ /analyzer /fieldType fieldtype name=phrase_suggest class=solr.TextField analyzer tokenizer class=solr.KeywordTokenizerFactory/ filter class=solr.PatternReplaceFilterFactory pattern=([^\p{L}\p{M}\p{N}\p{Cs}]*[\p{L}\p{M}\p{N}\p{Cs}\_]+:)|([^\p{L}\p{M}\p{N}\p{Cs}])+ replacement= replace=all/ filter class=solr.LowerCaseFilterFactory/ filter class=solr.TrimFilterFactory/ /analyzer /fieldtype fieldType name=text class=solr.TextField positionIncrementGap=100 analyzer type=index tokenizer class=solr.WhitespaceTokenizerFactory/ filter class=solr.StopFilterFactory ignoreCase=true words=stopwords.txt/ filter class=solr.WordDelimiterFilterFactory generateWordParts=1 generateNumberParts=1 catenateWords=1 catenateNumbers=1 catenateAll=0 splitOnCaseChange=1/ filter class=solr.LowerCaseFilterFactory/ filter class=solr.KeywordMarkerFilterFactory protected=protwords.txt/ filter class=solr.PorterStemFilterFactory/ filter class=solr.RemoveDuplicatesTokenFilterFactory/ /analyzer analyzer type=query tokenizer class=solr.WhitespaceTokenizerFactory/ filter class=solr.SynonymFilterFactory synonyms=synonyms.txt ignoreCase=true expand=true/ filter class=solr.StopFilterFactory ignoreCase=true words=stopwords.txt/ filter class=solr.WordDelimiterFilterFactory generateWordParts=1 generateNumberParts=1 catenateWords=0 catenateNumbers=0 catenateAll=0 splitOnCaseChange=1/ filter class=solr.LowerCaseFilterFactory/ filter class=solr.KeywordMarkerFilterFactory protected=protwords.txt/ filter class=solr.PorterStemFilterFactory/ filter class=solr.RemoveDuplicatesTokenFilterFactory/ /analyzer /fieldType field name=DisplayName type=text indexed=true stored=true required=true multiValued=false /
solrj getBeans from GroupCommand
I'm fairly new to Solr but think I have read the documentation well. Really enjoying the ease of using the likes of: ArrayListMyClass myRecords = queryResponse.getBeans(MyClass.class); However I have now come across the situation where I need to use groups. I cannot seem to find a way to getBeans() from a group, maybe I'm not understanding the hierarchy within a group? But I was hoping for something along the lines of: //Get beans from each group returnedList GroupCommand groupCommands = queryResponse.getGroupResponse().getValues();for(GroupCommand gc : groupCommands){ ListGroup groups = gc.getValues(); Group group = groups.get(0); ArrayListMyClass myRecords = group.getBeans(MyClass.class); //Do something with the current group of beans } Any advice or comments on how to do this or an alternative method would be gratefully received. Many thanks I have also posted this at: http://stackoverflow.com/questions/27114168/solrj-getbeans-from-groupcommand
Re: Backuping SolrCloud
Hi, There is a ticket for the same . https://issues.apache.org/jira/browse/SOLR-5750 Would you mind contributing to the discussion there? On Mon, Nov 24, 2014 at 11:07 AM, ralph tice ralph.t...@gmail.com wrote: I have a writeup of how to perform safe backups here: https://gist.github.com/ralph-tice/887414a7f8082a0cb828 There are some tickets around this work to further the ease of backups, especially https://issues.apache.org/jira/browse/SOLR-5750 On Mon, Nov 24, 2014 at 9:45 AM, Vivek Pathak vpat...@orgmeta.com wrote: I was able to get very good backup procedure by having cron job perform compact on each shard and then copy out the physical shard (the full lucene index directory to a backup server) Updates would need to be stopped during this period. And restore would be as simple as copying in the backed up shard and restarting solr On Nov 24, 2014, at 9:48 AM, elmerfudd na...@012.net.il wrote: Hi, I'm looking for a built-in SolrCloud backup mechanism. I want to backup my Index (scheduled / manual backups) while Indexing and searching. What is the proper way to perform this backup-restore task? Thanks. -- View this message in context: http://lucene.472066.n3.nabble.com/Backuping-SolrCloud-tp4170624.html Sent from the Solr - User mailing list archive at Nabble.com. -- - Noble Paul
Re: ERROR StreamingSolrServers 4.10.2
Hi Team, Does this mean that the updates on the other server is not getting forwarded to leader? Or can this error be safely ignored? ~Regards Joe On Mon, Nov 24, 2014 at 12:24 PM, Joseph V J josephveejay...@gmail.com wrote: Hi, I have a production Solr Cloud setup which has been migrated from 4.2 to 4.10.2. Upon then at times I'm getting this ERROR. ERROR StreamingSolrServers org.apache.solr.common.SolrException: Bad Request request: http://10.0.0.160:8080/solr/profiles/update?update.distrib=TOLEADERdistrib.from=http%3A%2F%2F10.0.2.160%3A8080%2Fsolr%2Fprofiles%2Fwt=javabinversion=2 at org.apache.solr.client.solrj.impl.ConcurrentUpdateSolrServer$Runner.run(ConcurrentUpdateSolrServer.java:241) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:745) The cloud is of single shard single replica setup with Zookeeper 3.4.6 and JAVA version used is java version 1.7.0_72 Java(TM) SE Runtime Environment (build 1.7.0_72-b14) Java HotSpot(TM) 64-Bit Server VM (build 24.72-b04, mixed mode) It would be great if anyone could throw some light here. ~Thanks Joe