Calculate a sum.
hello. My problem is, that i need to calculate a sum of amounts. this amount is in my index (stored=true). my php script get all values with paging. but if a request takes too long, jetty is killing this process and i get a broken pipe. Which is the best/fastest way to get the values of many fields from index? exists an ResponseHandler for exports? Or which is the fastest? -- View this message in context: http://lucene.472066.n3.nabble.com/Calculate-a-sum-tp4033091.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Calculate a sum.
Hello! Fetching all the documents, especially for a query that returns many documents can be a pain. However there is a StatsComponent (http://wiki.apache.org/solr/StatsComponent) in Solr, however your field would have to be numeric and indexed. -- Regards, Rafał Kuć Sematext :: http://sematext.com/ :: Solr - Lucene - Nutch - ElasticSearch hello. My problem is, that i need to calculate a sum of amounts. this amount is in my index (stored=true). my php script get all values with paging. but if a request takes too long, jetty is killing this process and i get a broken pipe. Which is the best/fastest way to get the values of many fields from index? exists an ResponseHandler for exports? Or which is the fastest? -- View this message in context: http://lucene.472066.n3.nabble.com/Calculate-a-sum-tp4033091.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Calculate a sum.
Hey, thx for your reply. i forgot to say. StatsComponent doesnt work with our application. too slow and buggy. but i test with this component with version 1.4 ... maybe some bugfixes in 4.0 ? this is the reason for calculating the sum on client side and some pages. but sometimes its too much for server. -- View this message in context: http://lucene.472066.n3.nabble.com/Calculate-a-sum-tp4033091p4033097.html Sent from the Solr - User mailing list archive at Nabble.com.
RE: CoreAdmin STATUS performance
Hi Stefan, I have opened issue SOLR-4302 and attached the suggested patch. Regards, Shahar. -Original Message- From: Stefan Matheis [mailto:matheis.ste...@gmail.com] Sent: Sunday, January 13, 2013 3:11 PM To: solr-user@lucene.apache.org Subject: Re: CoreAdmin STATUS performance Shahar would you mind, if i ask you to open an jira-issue for that? attaching your changes as typical patch? perhaps we could use that for the UI, in those cases where we don't need to full set of information .. Stefan On Sunday, January 13, 2013 at 12:28 PM, Shahar Davidson wrote: Shawn, Per and anyone else who has participated in this thread - thank you! I have finally resorted to apply a minor patch the Solr code. I have noticed that most of the time of the STATUS request is spent when collecting Index related info (such as segmentCount, sizeInBytes, numDocs, etc.). In the STATUS request I added support for a new parameter which, if present, will skip collection of the Index info (hence will only return general static info, among it the core name) - this will, in fact, cut down the request time by an order of two magnitudes! In my case, it decreased the request time from around 800ms to around 1ms-4ms. Regards, Shahar. -Original Message- From: Shawn Heisey [mailto:s...@elyograg.org] Sent: Thursday, January 10, 2013 5:14 PM To: solr-user@lucene.apache.org (mailto:solr-user@lucene.apache.org) Subject: Re: CoreAdmin STATUS performance On 1/10/2013 2:09 AM, Shahar Davidson wrote: As for your first question, the core info needs to be gathered upon every search request because cores are created dynamically. When a user initiates a search request, the system must be aware of all available cores in order to execute distributed search on _all_ relevant cores. (the user must get reliable and most up to date data) The reason that 800ms seems a lot to me is because the overall execution time takes about 2500ms and a large part of it is due to the STATUS request. The minimal interval concept is a good idea and indeed we've considered it, yet it poses a slight problem when building a RT system which needs to return to most up to date data. I am just trying to understand if there's some other way to hasten the STATUS reply (for example, by asking the STATUS request to return just certain core attributes, such as name, instead of collecting everything) Are there a *huge* number of SolrJ clients in the wild, or is it something like a server farm where you are in control of everything? If it's the latter, what I think I would do is have an asynchronous thread that periodically (every few seconds) updates the client's view of what cores exist. When a query is made, it will use that information, speeding up your queries by 800 milliseconds and ensuring that new cores will not have long delays before they become searchable. If you have a huge number of clients in the wild, it would still be possible, but ensuring that those clients get updated might be hard. If you also delete cores as well as add them, that complicates things. You'd have to have the clients be smart enough to exclude the last core on the list (by whatever sorting mechanism you require), and you'd have to wait long enough (30 seconds, maybe?) before *actually* deleting the last core to be sure that no clients are accessing it. Or you could use SolrCloud, as Per suggested, but with 4.1, not the released 4.0. SolrCloud manages your cores for you automatically. You'd probably be using a slightly customized SolrCloud, including the custom hashing capability added by SOLR-2592. I don't know what other customizations you might need. Thanks, Shawn Email secured by Check Point Email secured by Check Point
Re: Calculate a sum.
Stored fields are famous for its' slowness as well as they requires two io operation per doc. You can spend some heap for uninverting the index and utilize wiki.apache.org/solr/StatsComponent Let us know whether it works for you. 14.01.2013 13:14 пользователь stockii stock.jo...@googlemail.com написал: hello. My problem is, that i need to calculate a sum of amounts. this amount is in my index (stored=true). my php script get all values with paging. but if a request takes too long, jetty is killing this process and i get a broken pipe. Which is the best/fastest way to get the values of many fields from index? exists an ResponseHandler for exports? Or which is the fastest? -- View this message in context: http://lucene.472066.n3.nabble.com/Calculate-a-sum-tp4033091.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: theory of sets
Am 08.01.2013 10:26, schrieb Uwe Reh: OK, OK, I will try it again with dynamic fields. NO! dynamic fields are nice, but not for my problem. :-( I got more than *52* new fields. I was wrong, the impact on searching is really reasonable. But have you ever used the Admin's Schema Browser with that much fields? I suppose never, my Installation (4.1) freezes the FireFox while the JS-job runs into a timeout. At most I don't like it, because having that much fields 'smells' for me. Before anyone asks the XY question. The index is intended for a library's catalog and the quest is Find all members of a series (eg. penguin books, paperbacks) and order them on their sortkey. Unfortunately titles may belong to several (sub)series with different sortkeys. Still seeking for better approaches. Uwe
Re: configuring schema to match database
On 01/11/2013 06:14 PM, Gora Mohanty wrote: On 11 January 2013 22:30, Jens Grivolla j+...@grivolla.net wrote: [...] Actually, that is what you would get when doing a join in an RDBMS, the cross-product of your tables. This is NOT AT ALL what you typically do in Solr. Best start the other way around, think of Solr as a retrieval system, not a storage system. What are your queries? What do you want to find, and what criteria do you use to search for it? [...] Um, he did describe his desired queries, and there was a reason that I proposed the above schema design. He said he wants queries such as users how have taken courseA and are fluent in english, which is exactly one case I was describing. UserA has taken courseA, courseB and courseC and has writingskill good verbalskill good for english and writingskill excellent verbalskill excellent for spanish UserB has taken courseA, courseF, courseG and courseH and has writingskill fluent verbalskill fluent for english and writingskill good verbalskill good for italian Unless the index is becoming huge, I feel that it is better to flatten everything out rather than combine fields, and post-process the results. Then please show me the query to find users that are fluent in spanish and english. Bonus points if you manage to not retrieve the same user several times. (Hint, your schema stores only one language skill per row). Regards, Jens
Re: configuring schema to match database
On 14 January 2013 16:59, Jens Grivolla j+...@grivolla.net wrote: [...] Then please show me the query to find users that are fluent in spanish and english. Bonus points if you manage to not retrieve the same user several times. (Hint, your schema stores only one language skill per row). Doh! You are right, of course. Brainfart from my side. Regards, Gora
Re: Calculate a sum.
i've had perfectly fine performance with StatsComponent, but have only tested with 50,000 documents. for example i have field syllables and numeric field syllables_count. then i sum the syllable count for any search query. how many documents are you working with? On Mon, Jan 14, 2013 at 10:54 AM, Mikhail Khludnev mkhlud...@griddynamics.com wrote: Stored fields are famous for its' slowness as well as they requires two io operation per doc. You can spend some heap for uninverting the index and utilize wiki.apache.org/solr/StatsComponent Let us know whether it works for you. 14.01.2013 13:14 пользователь stockii stock.jo...@googlemail.com написал: hello. My problem is, that i need to calculate a sum of amounts. this amount is in my index (stored=true). my php script get all values with paging. but if a request takes too long, jetty is killing this process and i get a broken pipe. Which is the best/fastest way to get the values of many fields from index? exists an ResponseHandler for exports? Or which is the fastest? -- View this message in context: http://lucene.472066.n3.nabble.com/Calculate-a-sum-tp4033091.html Sent from the Solr - User mailing list archive at Nabble.com. -- edge
Re: configuring schema to match database
On 01/14/2013 12:50 PM, Gora Mohanty wrote: On 14 January 2013 16:59, Jens Grivolla j+...@grivolla.net wrote: [...] Then please show me the query to find users that are fluent in spanish and english. Bonus points if you manage to not retrieve the same user several times. (Hint, your schema stores only one language skill per row). Doh! You are right, of course. Brainfart from my side. Ok, I was starting to wonder if I was the one missing something. Re-reading what I wrote I see I may have sounded a bit rude, that was not my intention, sorry. Best, Jens
Re: How to manage solr cloud collections-sharding?
I can at least answer part of this see inline. On Sun, Jan 13, 2013 at 11:44 AM, adfel70 adfe...@gmail.com wrote: Hi, I know a few question on this issue have already been posted, but I dint find full answers in any of those posts. I'm using solr-4.0.0 [EOE] I'd _really_ start working with a nightly build instead, there have been a lot of improvements and the RC1 for 4.1 may well be cut this week. I need my solr cluster to have multiple collections, each collection with different configuration (at least different schema.xml file). I follow the solrCloud tutorial page and execute this command: /java -Dbootstrap_confdir=./solr/collection1/conf -Dcollection.configName=myconf -DzkRun -DnumShards=5 -jar start.jar/ when I start a solr servers I have collection1 in clustserState.json with each node assigned to some shard. questions so far: 1.Is this first command 100% necessary? [EOE] No, it's not. You could use the zkCli commands here: http://wiki.apache.org/solr/SolrCloud#Command_Line_Util. See especially the try bootstrapping all the conf dirs in solr.xml example. But sometime you have to send all the relevant info to Zookeeper, configuration files etc. This command is a convenience way to do that. 2. Do I have to defined the number of shards before starting solr instances? [EOE] Yes, unless you're doing custom sharding. 3. What if I want to add a shard after I started all solr instances and haven't indexed yet? [EOE] Then you have to re-index currently, unless you are doing custom hashing 4. what if I want to add a shard after indexing? [EOE] You have to re-index (and reconfigure your ZK state) unless you're doing custom sharding 5. what is the role that clustserState.json plays? is it just a json file to show in the GUI? Or is it the only file that persists the current state of the cluster? [EOE] What it does on ZK I don't know, but I've only seen it used as something for the GUI to read. Actually, all the other views are just prettifying this file. 6. Can I edit it manually? should I? [EOE] I've never heard of anyone even wanting to, you'd have to ask someone who knows way more about ZK than I do. I add another schema-B.xml file to the zookeeper and open another collection by using coreAdmin Rest API. I want this collection to have 10 shards and not 5 as I defined for the previous collection. So I run /http://server:port /solr/admin/cores?action=CREATEname=coreXinstanceDir=path_to_instance_directoryconfig=config_file_name.xmlschema=schem_file_name.xmldataDir=datashard=shard// 10 times with different / each run. [EOE] Currently, by adding the shard= parameter, you're now doing custom sharding. Mark just raised a JIRA about this recently, don't quite know what the current status of this is. You're in kind of uncharted territory here... questions: 1. is this an appropriate way to use the core admin API? should I specify the shard Id? I do it because it gives me a way to control the number of shards (each new shard id creates a new shard). but should I use it this way? [EOE] Right, but currently this means you do NOT get automatic distributed indexing, your indexing program has to send the document to the appropriate shard, preferably the leader. Like I said, this is kind of new. 2. Can I have different number of shards in different collections on the same cluster? 3. If yes - then what is the purpose of the first bootstrap command? [EOE] Well, without some kind of bootstrap, how would _any_ cluster information ever get to Zookeeper? The first bootstrap command is essentially the fire an forget approach so you don't have to keep track of anything to use SolrCloud. another question: I saw that in 4.1 version, each shard has another parameter - range. what is [EOE] Haven't worked with this yet this parameter used for? would I have to re-index when upgrading from 4.0 to 4.1? [EOE] You shouldn't have to re-index this will help a lot in understanding the whole collection-sharding architecture in solr cloud. Thanks -- View this message in context: http://lucene.472066.n3.nabble.com/How-to-manage-solr-cloud-collections-sharding-tp4033009.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: SolrCloud sort inconsistency
Unless it's a cut-n-paste error, you don't have an in front of the sort parameter so you're not sorting at all. You should have a sort section in your response where the params are echoed, something like: params:{ sort:id desc, fl:id, cache:False, q:id:*, rows:10}}, On Sun, Jan 13, 2013 at 6:42 PM, yriveiro yago.rive...@gmail.com wrote: How is possible that this sorted query returns different results? The highest value is the id P2450024023, sometimes the value returned is not the highest. This is an example, the second curl request is the correct result. NOTE: I did the query when a indexing process was running. ➜ ~ curl -H Cache-Control: no-cache http://192.168.1.241:8983/solr/ST-SHARD_0212/query\?q\=id :\*\rows\=10\fl\=id\=sort\=id%20desc\cache\=False { responseHeader:{ status:0, QTime:5, params:{ cache:False, rows:10, fl:id=sort=id desc, q:id:*}}, response:{numFound:2387312,start:0,maxScore:1.0,docs:[ { id:P2443605077}, { id:P2443588094}, { id:P2443647855}, { id:P2443613193}, { id:P2443572098}, { id:P2443562507}, { id:P2443643935}, { id:P2443556464}, { id:P2443625267}, { id:P2443580781}] }} ➜ ~ curl -H Cache-Control: no-cache http://192.168.1.241:8983/solr/ST-SHARD_0212/query\?q\=id :\*\rows\=10\fl\=id\=sort\=id%20desc\cache\=False { responseHeader:{ status:0, QTime:4, params:{ cache:False, rows:10, fl:id=sort=id desc, q:id:*}}, response:{numFound:2387312,start:0,maxScore:1.0,docs:[ { id:P2450024023}, { id:P2450017490}, { id:P2450062568}, { id:P2450053498}, { id:P2449990839}, { id:P2449973572}, { id:P2449957535}, { id:P2450099098}, { id:P2450090195}, { id:P2450072528}] }} ➜ ~ curl -H Cache-Control: no-cache http://192.168.1.241:8983/solr/ST-SHARD_0212/query\?q\=id :\*\rows\=10\fl\=id\=sort\=id%20desc\cache\=False { responseHeader:{ status:0, QTime:6, params:{ cache:False, rows:10, fl:id=sort=id desc, q:id:*}}, response:{numFound:2387312,start:0,maxScore:1.0,docs:[ { id:P2450024023}, { id:P2450017490}, { id:P2450062568}, { id:P2450053498}, { id:P2449990839}, { id:P2449973572}, { id:P2449957535}, { id:P2450099098}, { id:P2450090195}, { id:P2450072528}] }} ➜ ~ - Best regards -- View this message in context: http://lucene.472066.n3.nabble.com/SolrCloud-sort-inconsistency-tp4033046.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: SolrCloud sort inconsistency
P.S. of course your sorting won't reflect documents that haven't been committed yet. So if you straighten out the params your lists should be in order, but the documents returned may change depending on whether your indexing process adds docs between calls Erick On Mon, Jan 14, 2013 at 7:25 AM, Erick Erickson erickerick...@gmail.comwrote: Unless it's a cut-n-paste error, you don't have an in front of the sort parameter so you're not sorting at all. You should have a sort section in your response where the params are echoed, something like: params:{ sort:id desc, fl:id, cache:False, q:id:*, rows:10}}, On Sun, Jan 13, 2013 at 6:42 PM, yriveiro yago.rive...@gmail.com wrote: How is possible that this sorted query returns different results? The highest value is the id P2450024023, sometimes the value returned is not the highest. This is an example, the second curl request is the correct result. NOTE: I did the query when a indexing process was running. ➜ ~ curl -H Cache-Control: no-cache http://192.168.1.241:8983/solr/ST-SHARD_0212/query\?q\=id :\*\rows\=10\fl\=id\=sort\=id%20desc\cache\=False { responseHeader:{ status:0, QTime:5, params:{ cache:False, rows:10, fl:id=sort=id desc, q:id:*}}, response:{numFound:2387312,start:0,maxScore:1.0,docs:[ { id:P2443605077}, { id:P2443588094}, { id:P2443647855}, { id:P2443613193}, { id:P2443572098}, { id:P2443562507}, { id:P2443643935}, { id:P2443556464}, { id:P2443625267}, { id:P2443580781}] }} ➜ ~ curl -H Cache-Control: no-cache http://192.168.1.241:8983/solr/ST-SHARD_0212/query\?q\=id :\*\rows\=10\fl\=id\=sort\=id%20desc\cache\=False { responseHeader:{ status:0, QTime:4, params:{ cache:False, rows:10, fl:id=sort=id desc, q:id:*}}, response:{numFound:2387312,start:0,maxScore:1.0,docs:[ { id:P2450024023}, { id:P2450017490}, { id:P2450062568}, { id:P2450053498}, { id:P2449990839}, { id:P2449973572}, { id:P2449957535}, { id:P2450099098}, { id:P2450090195}, { id:P2450072528}] }} ➜ ~ curl -H Cache-Control: no-cache http://192.168.1.241:8983/solr/ST-SHARD_0212/query\?q\=id :\*\rows\=10\fl\=id\=sort\=id%20desc\cache\=False { responseHeader:{ status:0, QTime:6, params:{ cache:False, rows:10, fl:id=sort=id desc, q:id:*}}, response:{numFound:2387312,start:0,maxScore:1.0,docs:[ { id:P2450024023}, { id:P2450017490}, { id:P2450062568}, { id:P2450053498}, { id:P2449990839}, { id:P2449973572}, { id:P2449957535}, { id:P2450099098}, { id:P2450090195}, { id:P2450072528}] }} ➜ ~ - Best regards -- View this message in context: http://lucene.472066.n3.nabble.com/SolrCloud-sort-inconsistency-tp4033046.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: theory of sets
Does this have to be in Solr? Given the pre-computed nature of the sub-series, maybe you can encode both series name and series sort order in a separate structure designed for it. Something like Neo4J comes to mind: http://www.neo4j.org/ . Or, this could be a good question for StackOverflow on what handles this kind of scenarios best. Regards, Alex. On Mon, Jan 14, 2013 at 5:55 AM, Uwe Reh r...@hebis.uni-frankfurt.de wrote: Before anyone asks the XY question. The index is intended for a library's catalog and the quest is Find all members of a series (eg. penguin books, paperbacks) and order them on their sortkey. Unfortunately titles may belong to several (sub)series with different sortkeys. Still seeking for better approaches. Personal blog: http://blog.outerthoughts.com/ LinkedIn: http://www.linkedin.com/in/alexandrerafalovitch - Time is the quality of nature that keeps events from happening all at once. Lately, it doesn't seem to be working. (Anonymous - via GTD book)
Re: Calculate a sum.
Mikhail Khludnev wrote You can spend some heap for uninverting the index and utilize wiki.apache.org/solr/StatsComponent what do you mean with this? Edward Garrett wrote how many documents are you working with? ~90 million documents ... -- View this message in context: http://lucene.472066.n3.nabble.com/Calculate-a-sum-tp4033091p4033152.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: regex and highlighter component: highlight and return individual fragments inside a snippet
it seems, hl.snippets does what I'm after, also discussed here: http://search-lucene.com/m/yxCOc1X0uY42/highlighting+multiple+occurrencessubj=highlighting+multiple+occurrences http://wiki.apache.org/solr/*Highlighting*Parameters#hl.snippetshttp://wiki.apache.org/solr/HighlightingParameters#hl.snippets Dmitry On Mon, Jan 14, 2013 at 3:14 PM, Dmitry Kan solrexp...@gmail.com wrote: Hello! I'm playing with the regex feature of highlighting in SOLR. The regex I have is pretty simple and, given a keyword query, it hits in a few places inside each document. Is there a way of highlighting and returning individual fragments with this approach? That is, if the matches are found let's say in the beginning, middle and end parts of the document, then return only these three groups inside a highlighter snippet? If there isn't such a functionality, do you know, what would it take to implement one (solr/lucene 3.4)? Regards, Dmitry Kan
How to use shardId
Hi all, I am trying to realize a solr cloud cluster with 2 collections and 4 shards each with 2 replicates hosted by 4 solr instances. If shardNum parm is set to 4 and all solr instances are started after each other it seems to work fine. What I wanted to do now is removing shardNum from JAVA_OPTS and defining each core with a shardId. Here is my current solr.xml of the first and second (in the second there is another instanceDir, the rest is the same) solr instance: Here is solr.xml of the third and fourth solr instance: But it seems that solr doesn't accept the shardId or omits it. What I really get is 2 collections each with 2 shards and 8 replicates (each solr instance 2) Either the functionality is not really clear to me or there has to be a config failure. It would very helpful if anyone could give me a hint. Thanks. starbuck -- View this message in context: http://lucene.472066.n3.nabble.com/How-to-use-shardId-tp4033186.html Sent from the Solr - User mailing list archive at Nabble.com.
RE: SOlr 3.5 and sharding
If you have the same documents -- with the same uniqueKey -- across multiple shards, the count will not be what you expect. You'll need to ensure that each document exists only on a single shard. -Michael -Original Message- From: Jean-Sebastien Vachon [mailto:jean-sebastien.vac...@wantedanalytics.com] Sent: Monday, January 14, 2013 9:59 AM To: solr-user@lucene.apache.org Subject: SOlr 3.5 and sharding Hi, I`m setting up a small Sorl setup consisting of 1 master node and 4 shards. For now, all four shards contains the exact same data. When I perform a query on each individual shards for the word `java` I am receiving the same number of docs (as expected). However, when I am going through the master node using the shards parameters, the number of results is slightly off by a few documents. There is nothing special in my setup so I`m looking for hints on why I am getting this problem Thanks
Re: RSS tutorial that comes with the apache-solr not indexing
This example may be out of date, if the RSS feeds from Slashdot have changed. If you know XML and XPaths, try this: Find an rss feed from somewhere that works. Compare the xpaths in it v.s. the xpaths in the DIH script. On 01/13/2013 07:38 PM, bibhor wrote: Hi I am trying to use the RSS tutorial that comes with the apache-solr. I am not sure if I missed anything but when I do full-import no indexing happens. These are the steps that I am taking: 1) Download apache-solr-3.6.2 (http://lucene.apache.org/solr/) 2) Start the solr by doing: java -Dsolr.solr.home=./example-DIH/solr/ -jar start.jar 3) Goto url: http://192.168.1.12:8983/solr/rss/dataimport?command=full-import 4) When I do this it says: Indexing completed. Added/Updated: 0 documents. Deleted 0 documents. Now I know that the default example is getting the RSS from: http://rss.slashdot.org/Slashdot/slashdot This default example is empty when I view it in chrome. It does have XML data in the source but I am not sure if this has anything to do with the import failure. I also modified the rss-config so that I can test other RSS sources. I used http://www.feedforall.com/sample.xml and updated the rss-config.xml but this did the same and did not Add/Update any documents. Any help is appreciated. -- View this message in context: http://lucene.472066.n3.nabble.com/RSS-tutorial-that-comes-with-the-apache-solr-not-indexing-tp4033067.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Schema Field Names i18n
Will a field have different names in different languages? There is no facility for 'aliases' for field name. Erick is right, this sounds like you need query and update components to implement this. Also, you might try using URL-encoding for the field names. This would save my sanity. On 01/10/2013 04:56 AM, Erick Erickson wrote: There's no really easy way that I know of. I've seen several approaches used though 1 do it in the UI. This assumes that your users aren't typing in raw queries, they're picking field names from a drop-down or similar. Then the UI maps the chosen fields into what the schema defines. 2 Do it in the middleware when assembling the query to pass through. Be careful with the translations though, there always seem to be edge cases. 3 What you're suggesting. Unless you're really fluent in parsers (they give me indigestion) I'd think about a query component. Best Erick On Wed, Jan 9, 2013 at 7:36 PM, Daryl Robbins daryl.robb...@mxi.com wrote: Anyone have experience with internationalizing the field names in the SOLR schema, so users in different languages can specify fields in their own language? My first thoughts would be to create a custom search component or query parser than would convert localized field names back to the English names in the schema, but I haven't dived in too deep yet. Any input would be greatly appreciated. Thanks, Daryl __ * This message is intended only for the use of the individual or entity to which it is addressed, and may contain information that is privileged, confidential and exempt from disclosure under applicable law. Unless you are the addressee (or authorized to receive for the addressee), you may not use, copy or disclose the message or any information contained in the message. If you have received this message in error, please advise the sender by reply e-mail, and delete the message, or call +1-613-747-4698. *
Re: Solr 4.0 SnapPuller version vs. generation issue
I've fixed this - thanks Gregg. https://issues.apache.org/jira/browse/SOLR-4303 - Mark On Jan 10, 2013, at 5:41 PM, Mark Miller markrmil...@gmail.com wrote: Hmm…I don't recall that change. We use the force, so SolrCloud certainly does not depend on it. It seems like it might be a mistake - some dev code that got caught up with the commit? I'm a little surprised it wouldn't trip any tests…I still have to read your first email closely though. - Mark On Jan 10, 2013, at 4:49 PM, Gregg Donovan gregg...@gmail.com wrote: Thanks, Mark. The relevant commit on the solrcloud branch appears to be 1231134 and is focused on the recovery aspect of SolrCloud: http://svn.apache.org/viewvc?diff_format=hview=revisionrevision=1231134 http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/solr/core/src/java/org/apache/solr/handler/SnapPuller.java?diff_format=hr1=1231133r2=1231134; I tried changed the check on our 4.0 test cluster to: boolean isFullCopyNeeded = IndexDeletionPolicyWrapper.getCommitTimestamp(commit) = latestVersion || commit.getGeneration() = latestGeneration || forceReplication; and that fixed our post-reindexing HTTP replication issues. But I'm not sure if that check works for all of the cases that SnapPuller is designed for. --Gregg On Thu, Jan 10, 2013 at 4:28 PM, Mark Miller markrmil...@gmail.com wrote: On Jan 10, 2013, at 4:11 PM, Gregg Donovan gregg...@gmail.com wrote: If the commitTimeMSec based check in Solr 4.0 is needed for SolrCloud, It's not. SolrCloud just uses the force option. I think this other change was made because Lucene stopped using both generation and version. I can try and look closer later - can't remember who made the change in Solr. - Mark
Re: RSS tutorial that comes with the apache-solr not indexing
Hi, I did try another RSS from here http://www.feedforall.com/sample.xml; but it also didnt work and came back with same message saying it indexed 0 documents. This is my data from rss-data-config.xml dataConfig dataSource type=URLDataSource / document entity name=rsstest pk=link url=http://www.feedforall.com/sample.xml; processor=XPathEntityProcessor forEach=/RDF/channel | /RDF/item transformer=DateFormatTransformer field column=title xpath=/RDF/item/title / field column=link xpath=/RDF/item/link / /entity /document /dataConfig -- View this message in context: http://lucene.472066.n3.nabble.com/RSS-tutorial-that-comes-with-the-apache-solr-not-indexing-tp4033067p4033230.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: RSS tutorial that comes with the apache-solr not indexing
Hi bibhor, I looked at http://rss.slashdot.org/Slashdot/slashdot and http://www.feedforall.com/sample.xml, and their top-level structure is: rss channel ... This doesn't match your entity ... forEach=/RDF/channel ... or your field column=... xpath=/RDF/item/... / Steve On Jan 14, 2013, at 1:02 PM, bibhor bib...@gmail.com wrote: Hi, I did try another RSS from here http://www.feedforall.com/sample.xml; but it also didnt work and came back with same message saying it indexed 0 documents. This is my data from rss-data-config.xml dataConfig dataSource type=URLDataSource / document entity name=rsstest pk=link url=http://www.feedforall.com/sample.xml; processor=XPathEntityProcessor forEach=/RDF/channel | /RDF/item transformer=DateFormatTransformer field column=title xpath=/RDF/item/title / field column=link xpath=/RDF/item/link / /entity /document /dataConfig -- View this message in context: http://lucene.472066.n3.nabble.com/RSS-tutorial-that-comes-with-the-apache-solr-not-indexing-tp4033067p4033230.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: configuring schema to match database
On 14 January 2013 17:28, Jens Grivolla j+...@grivolla.net wrote: On 01/14/2013 12:50 PM, Gora Mohanty wrote: [...] Doh! You are right, of course. Brainfart from my side. Ok, I was starting to wonder if I was the one missing something. Re-reading what I wrote I see I may have sounded a bit rude, that was not my intention, sorry. Did not take it as rude, and in any case am willing to tolerate a lot of impoliteness when someone shows me that I was wrong. Must have been half-asleep when I wrote my original reply, and was then trying to defend it. At least that's my story, and I am sticking to it :-) Regards, Gora
Re: RSS tutorial that comes with the apache-solr not indexing
Hi Steve Thank you for your help. After I updated the rss-data-config to following, it worked. dataConfig dataSource type=URLDataSource / document entity name=rsstest pk=link url=http://www.feedforall.com/sample.xml; processor=XPathEntityProcessor forEach=/rss/channel/item transformer=DateFormatTransformer field column=title xpath=/rss/channel/item/title / field column=link xpath=/rss/channel/item/link / /entity /document /dataConfig -- View this message in context: http://lucene.472066.n3.nabble.com/RSS-tutorial-that-comes-with-the-apache-solr-not-indexing-tp4033067p4033254.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: performance improvements on ip look up query
Hello Lee, I suppose caching isn't efficient for this type of searches. I can propose a kind of trick. if you index your docs in order by (STARTIP,ENDIP) tuple, it should make intersection faster. However it's rather theoretical consideration than practical one. More real one is to encode ip range in a single field and search by special query. I'm aware of geospatial approach for this lucene.472066.n3.nabble.com/Modeling-openinghours-using-multipoints-td4025336.html Unfortunately I never did it. On Wed, Jan 9, 2013 at 8:13 PM, Lee Carroll lee.a.carr...@googlemail.comwrote: Hi Otis The cache was modest 4096 with a hit rate of 0.23 after a 24hr period. We doubled it and the hit rate went to 0.25. Our interpretation is ip is pretty much a cache busting value ? and cache size is not at play here. the q param is just startIpNum:[* TO 180891652]AND endIpNum:[180891652 TO *] so again our interpretation is its got little reuse Could we re-formulate the query to be more per-formant ? On 9 January 2013 12:56, Otis Gospodnetic otis.gospodne...@gmail.com wrote: Hi, Maybe your cache is too small? How big is it and does the hit rate change if you make it bigger? Do any parts of the query repeat a lot? Maybe there is room for fq. Otis Solr ElasticSearch Support http://sematext.com/ On Jan 9, 2013 6:08 AM, Lee Carroll lee.a.carr...@googlemail.com wrote: Hi We are doing a lat/lon look up query using ip address. We have a 6.5 million document core of the following structure start ip block end ip block location id location_lat_lon the field defs are types fieldType name=string class=solr.StrField sortMissingLast=true omitNorms=true/ fieldType name=tlong class=solr.TrieLongField precisionStep=8 omitNorms=true positionIncrementGap=0/ fieldType name=tfloat class=solr.TrieFloatField precisionStep=8 omitNorms=true positionIncrementGap=0/ fieldType name=tdouble class=solr.TrieDoubleField precisionStep=8 omitNorms=true positionIncrementGap=0/ fieldType name=location class=solr.LatLonType subFieldSuffix= _coordinate/ /types fields field name=startIp type=string indexed=true stored=false required= true/ field name=startIpNum type=tlong indexed=true stored=false required =true/ field name=endIpNum type=tlong indexed=true stored=false required= true/ field name=locId type=string indexed=true stored=true required= true/ field name=countryCode type=string indexed=true stored=true required=false/ field name=cityName type=string indexed=false stored=true required =false/ field name=latLon type=location indexed=true stored=true required= true/ field name=latitude type=string indexed=false stored=true required =true/ field name=longitude type=string indexed=false stored=true required =true/ dynamicField name=*_coordinate type=tdouble indexed=true stored= false/ /fields the query at the moment is simply a range query q=startIpNum:[* TO 180891652]AND endIpNum:[180891652 TO *] we are seeing a full query cache with a low hit rate 0.2 and a high eviction rate which makes sense given the use of ip address in the query. query time mean is 120. Is their a better way of structuring the core for this usecase ? I suspect our heap memory settings are conservative 1g but will need to convince our sys admins to change this (they are not ringing any resource alarm bells) just the query is a little slow -- Sincerely yours Mikhail Khludnev Principal Engineer, Grid Dynamics http://www.griddynamics.com mkhlud...@griddynamics.com
Velocity in Multicore
Hi, I've been unable to get the velocity response writer to work in a multicore environment. Working from the examples that are distributed with Solr I simply started from the multicore example and added a hello.vm into core0/conf/velocity directory. I then updated the solrconfig.xml to add a new request handler as shown below. I've tried to use the v.base_dir to no success. Essentially what I always end up with is the default solr response. Has anyone been able to get the velocity response writer to work in a multicore environment? If so, could you point me to the documentation on how to do so. hello.vm Hello World! solrconfig.xml === … requestHandler name=/hello class=solr.SearchHandler lst name=defaults str name=echoParamsexplicit/str !-- VelocityResponseWriter settings -- str name=wtvelocity/str str name=v.templatehello/str !-- I've tried all the following in addition to not specifying any. -- !--str name=v.base_dircore0/conf/velocity/str-- !--str name=v.base_dirconf/velocity/str-- !--str name=v.base_dirmulticore/core0/conf/velocity/str-- /lst /requestHandler … Regards, Paul Ramirez
Re: POST query with non-ASCII to solr using httpclient wont work
unfortunately solrj is not an option here... we will have to make a quick fix with a patch out in production. I am still unable to make the solr (3.5) take url encoded query. again passing non-urlencoded query string works with non-ASIIC (Chinese), but fails return anything when sending request with urlencoded + Chinese. any suggestion? thanks jie -- View this message in context: http://lucene.472066.n3.nabble.com/POST-query-with-non-ASCII-to-solr-using-httpclient-wont-work-tp4032957p4033262.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: RSS tutorial that comes with the apache-solr not indexing
On 15 January 2013 00:38, Alexandre Rafalovitch arafa...@gmail.com wrote: Is that something that needs to be updated in the example schema as well then? The example rss-data-config.xml references http://rss.slashdot.org/Slashdot/slashdot which seems to be broken at the moment, at least for me. This is the same URL as on the Slashdot home page, so not sure what is going on. Regards, Gora
Re: theory of sets
My answer as usual - BlockJoin. index group as a parent document, and every membership as a child doc. In this case you somehow denormalize your items - every item will be indexed N times where N is a number of groups it belongs to. Potentially it can lead to duplication problem, but you haven't mention it yet. In this case single field for sorting is enough. then you can search by ToChildQuery(GRP_NAME:foo) than gives you foo's children. BJQ-101: http://blog.mikemccandless.com/2012/01/searching-relational-content-with.html http://blog.mikemccandless.com/2012/01/tochildblockjoinquery-in-lucene.html https://issues.apache.org/jira/browse/SOLR-3076 forgot to mention there is a price for blazing speed, when you need to change membership you need to reindex whole block. On Fri, Jan 4, 2013 at 1:10 AM, Uwe Reh r...@hebis.uni-frankfurt.de wrote: Hi, I'm looking for a tricky solution of a common problem. I have to handle a lot of items and each could be member of several groups. - OK, just add a field called 'member_of' No that's not enough, because each group is sorted and each member has a sortstring for this group. - OK, still easy add a dynamic field 'sortinfo_for_*' and fill this for each group membership. Yes, this works, but there are thousands of different groups, that much dynamic fields are probably a serious performance issue. - Well ... I'm looking for a smart way to answer to the question Find the members of group X and sort them by the the sortstring for this group. One idea I had was to fill the 'member_of' field with composed entrys (groupname + _ + sortstring). Finding the members is easy with wildcards but there seems to be no way to use the sortstring as a boostfactor Has anybody solved this problem? Any hints are welcome. Uwe -- Sincerely yours Mikhail Khludnev Principal Engineer, Grid Dynamics http://www.griddynamics.com mkhlud...@griddynamics.com
Re: Search across a specified number of boundaries
Mike, When Lucene's Analyser indexes the text it adds positions into the index which are lately used by SpanQueries. Have you considered idea of position increment gap? e.g. the first sentence is indexed with words positions: 0,1,2,3,... the second sentence with 100,101,102,103,..., third 200,201,202.. Then applying some span constraint allows you search across/inside of the sentences. WDYT? On Sun, Jan 6, 2013 at 6:50 PM, Erick Erickson erickerick...@gmail.comwrote: Mike: I'm _really_ stretching here, but you might be able to do something interesting with payloads. Say each word had a payload with the sentence number and you _somehow_ made use of that information in a custom scorer. But like I said, I really have no good idea how to accomplish that... BTW, in future this kind of question is better asked on the user's list (either Lucene or Solr), this list if intended for discussing development work Best Erick On Fri, Jan 4, 2013 at 1:02 PM, Mike Ree mike.ad...@olytech.net wrote: d terms that are in nearby sentences. IE: TermA NEAR3 TermB would find all TermA's that are within 3 sentences of TermB. Have found ways to find TermA within same sentence -- Sincerely yours Mikhail Khludnev Principal Engineer, Grid Dynamics http://www.griddynamics.com mkhlud...@griddynamics.com
Re: Question about GC logging timestamps
Shawn, you are welcome! http://www.oracle.com/technetwork/java/gc-tuning-5-138395.html The flag -XX:+PrintGCTimeStamps will additionally print a time stamp at the start of each collection. On Sun, Jan 6, 2013 at 6:54 AM, Michael Ryan mr...@moreover.com wrote: From my own experience, the timestamp seems to be logged at the start of the garbage collection. -Michael -- Sincerely yours Mikhail Khludnev Principal Engineer, Grid Dynamics http://www.griddynamics.com mkhlud...@griddynamics.com
Re: Index data from multiple tables into Solr
Try all of the links under the collection name in the lower left-hand columns. There several administration monitoring tools you may find useful. On 01/14/2013 11:45 AM, hassancrowdc wrote: ok stats are changing, so the data is indexed. But how can i do query with this data, or ow can i search it, like the command will be http://localhost:8983/solr/select?q=(any of my field column from table)? coz whatever i am putting in my url it shows me an xml file but the numFound are always 0? On Sat, Jan 12, 2013 at 1:24 PM, Alexandre Rafalovitch [via Lucene] ml-node+s472066n4032778...@n3.nabble.com wrote: Have you tried the Admin interface yet? The one on :8983 port if you are running default setup. That has a bunch of different stats you can look at apart from a nice way of doing a query. I am assuming you are on Solr 4, of course. Regards, Alex. On Fri, Jan 11, 2013 at 5:13 PM, hassancrowdc [hidden email]http://user/SendEmail.jtp?type=nodenode=4032778i=0wrote: So, I followed all the steps and solr is working successfully, Can you please tell me how i can see if my data is indexed or not? do i have to enter specific url into my browser or anything. I want to make sure that the data is indexed. Personal blog: http://blog.outerthoughts.com/ LinkedIn: http://www.linkedin.com/in/alexandrerafalovitch - Time is the quality of nature that keeps events from happening all at once. Lately, it doesn't seem to be working. (Anonymous - via GTD book) -- If you reply to this email, your message will be added to the discussion below: http://lucene.472066.n3.nabble.com/Index-data-from-multiple-tables-into-Solr-tp4032266p4032778.html To unsubscribe from Index data from multiple tables into Solr, click herehttp://lucene.472066.n3.nabble.com/template/NamlServlet.jtp?macro=unsubscribe_by_codenode=4032266code=aGFzc2FuY3Jvd2RjYXJlQGdtYWlsLmNvbXw0MDMyMjY2fC00ODMwNzMyOTM= . NAMLhttp://lucene.472066.n3.nabble.com/template/NamlServlet.jtp?macro=macro_viewerid=instant_html%21nabble%3Aemail.namlbase=nabble.naml.namespaces.BasicNamespace-nabble.view.web.template.NabbleNamespace-nabble.view.web.template.NodeNamespacebreadcrumbs=notify_subscribers%21nabble%3Aemail.naml-instant_emails%21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml -- View this message in context: http://lucene.472066.n3.nabble.com/Index-data-from-multiple-tables-into-Solr-tp4032266p4033268.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: POST query with non-ASCII to solr using httpclient wont work
Hi Jie, maybe there is a simple solution. When we used tomcat as servlet container for solr I notices similar problems. Even with the hints from the solr wiki about unicode and Tomcat, i wasn't able to fix this. So we switched back to Jetty, querys like q=allfields2%3A能力 are reliable now. Uwe BTW: I have no idea for at all what these Japanese signs mean. So just let me append two of 31 hits in our bibliographic catalog doc str name=idHEB052032124/str str name=raw_fullrecordalg: 5203212 001@ $0205 001A $4:13-05-97 001B $t13:12:07.000$01999:10-06-10 001D $0:99-99-99 001U $0utf8 001X $00 002@ $0Aau 003@ $0052032124 007I $0NacsisBN09679884 010@ $ajpn 011@ $a1993 013H $0z 019@ $ajp 021A $ULatn$T01$aNōryoku kaihatsu no shisutemu$hYaguchi Hajime 021A $UJpan$T01$a@能力開発のシステム$h矢口新著 028A $ULatn$T01$9165745363$8Yaguchi, Hajime 028A $UJpan$T01$d新$a矢口 033A $ULatn$T01$pTokyo$nNōryoku Kaihatsu Kōgaku Sentaa 033A $UJpan$T01$p東久留米$n能力開発工学センター 034D $a274 S. 034M $aIll. 036E $aYaguchi Hajime senshū$l2 036F $l2$9052031527$8Yaguchi Hajime senshū$x12 037B $aSysteme zur Entwicklung der Fähigkeiten 046L $aIn japan. Schr. ... 247C/01 $9102595631$8351457-2 4/457Marburg, Universität Marburg, Bibliothek des Japan-Zentrums (BJZ) /str /doc doc str name=idHEB286840723/str str name=raw_fullrecordalg: 28684072 001@ $03 001A $00030:04-01-12 001B $t22:29:11.000$01999:04-01-12 001C $t10:48:47.000$00030:04-01-12 001D $00030:04-01-12 001U $0utf8 001X $00 002@ $0Aau 003@ $0286840723 004A $A978-4-88319-546-6 007A $0286840723$aHEB 010@ $ajpn 011@ $a2010 021A $ULatn$T01$aShin kanzen masutā kanji nihongo nōryoku shiken ; N1$hIshii Reiko ... 021A $UJpan$T01$a新完全マスター漢字日本語能力試験 ; N1$h石井怜子 [ほか] 著 027A $ULatn$T01$aShin kanzen masutā kanji : nihongo nōryoku shiken ; enu ichi / Ishii Reiko ... 027A $UJpan$T01$a新完全マスター漢字 : 日本語能力試験 ; N1 / 石井怜子 [ほか] 著 028C $9230917593$8Ishii, Reiko 033A $ULatn$T01$pTōkyō$nSurīē nettowāku 033A $UJpan$T01$p東京$nスリーエーネットワーク 034D $aviii, 197, 21S. 034I $a26cm 044A $S4$aNihongokyōiku(Taigaikokujin) 045Z $aEI 4650 ... 247C/01 $9102599157$8601220-6 30/220Frankfurt, Universität Frankfurt, Institut für Orientalische und Ostasiatische Philologien, Japanologie /str /doc
Re: Index data from multiple tables into Solr
thanx, I got it. How Can i integrate solr with my website? so that i can use it for search? On Mon, Jan 14, 2013 at 4:04 PM, Lance Norskog-2 [via Lucene] ml-node+s472066n4033291...@n3.nabble.com wrote: Try all of the links under the collection name in the lower left-hand columns. There several administration monitoring tools you may find useful. On 01/14/2013 11:45 AM, hassancrowdc wrote: ok stats are changing, so the data is indexed. But how can i do query with this data, or ow can i search it, like the command will be http://localhost:8983/solr/select?q=(any of my field column from table)? coz whatever i am putting in my url it shows me an xml file but the numFound are always 0? On Sat, Jan 12, 2013 at 1:24 PM, Alexandre Rafalovitch [via Lucene] [hidden email] http://user/SendEmail.jtp?type=nodenode=4033291i=0 wrote: Have you tried the Admin interface yet? The one on :8983 port if you are running default setup. That has a bunch of different stats you can look at apart from a nice way of doing a query. I am assuming you are on Solr 4, of course. Regards, Alex. On Fri, Jan 11, 2013 at 5:13 PM, hassancrowdc [hidden email] http://user/SendEmail.jtp?type=nodenode=4032778i=0wrote: So, I followed all the steps and solr is working successfully, Can you please tell me how i can see if my data is indexed or not? do i have to enter specific url into my browser or anything. I want to make sure that the data is indexed. Personal blog: http://blog.outerthoughts.com/ LinkedIn: http://www.linkedin.com/in/alexandrerafalovitch - Time is the quality of nature that keeps events from happening all at once. Lately, it doesn't seem to be working. (Anonymous - via GTD book) -- If you reply to this email, your message will be added to the discussion below: . NAML http://lucene.472066.n3.nabble.com/template/NamlServlet.jtp?macro=macro_viewerid=instant_html%21nabble%3Aemail.namlbase=nabble.naml.namespaces.BasicNamespace-nabble.view.web.template.NabbleNamespace-nabble.view.web.template.NodeNamespacebreadcrumbs=notify_subscribers%21nabble%3Aemail.naml-instant_emails%21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml -- View this message in context: http://lucene.472066.n3.nabble.com/Index-data-from-multiple-tables-into-Solr-tp4032266p4033268.html Sent from the Solr - User mailing list archive at Nabble.com. -- If you reply to this email, your message will be added to the discussion below: http://lucene.472066.n3.nabble.com/Index-data-from-multiple-tables-into-Solr-tp4032266p4033291.html To unsubscribe from Index data from multiple tables into Solr, click herehttp://lucene.472066.n3.nabble.com/template/NamlServlet.jtp?macro=unsubscribe_by_codenode=4032266code=aGFzc2FuY3Jvd2RjYXJlQGdtYWlsLmNvbXw0MDMyMjY2fC00ODMwNzMyOTM= . NAMLhttp://lucene.472066.n3.nabble.com/template/NamlServlet.jtp?macro=macro_viewerid=instant_html%21nabble%3Aemail.namlbase=nabble.naml.namespaces.BasicNamespace-nabble.view.web.template.NabbleNamespace-nabble.view.web.template.NodeNamespacebreadcrumbs=notify_subscribers%21nabble%3Aemail.naml-instant_emails%21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml -- View this message in context: http://lucene.472066.n3.nabble.com/Index-data-from-multiple-tables-into-Solr-tp4032266p4033296.html Sent from the Solr - User mailing list archive at Nabble.com.
SolrCloud :: Adding replica :: Sync-up issue
Hello, I observed a rather weird issue with SolrCloud. Using Solr 4.0 GA code. Started with a 3-node Zookeeper ensemble (standalone) and a single Solr instance running single collection. numShards was set to 1 during collection creation (don't want sharding, just replication). Everything worked fine. Started another Solr instance for the same collection. Properly went through the steps realizing it needed to sync up (actual url values replaced with url): 12:50:59.152 INFO [o.apache.solr.cloud.RecoveryStrategy] Starting recovery process. core=users recoveringAfterStartup=true [RecoveryThread] 12:50:59.152 INFO [o.a.solr.servlet.SolrDispatchFilter ] user.dir=/home/seg [localhost-startStop-1] 12:50:59.153 INFO [o.a.solr.servlet.SolrDispatchFilter ] SolrDispatchFilter.init() done [localhost-startStop-1] 12:50:59.189 INFO [o.apache.solr.cloud.RecoveryStrategy] ## startupVersions=[] [RecoveryThread] 12:50:59.198 INFO [o.apache.solr.cloud.RecoveryStrategy] Attempting to PeerSync from url core=users - recoveringAfterStartup=true [RecoveryThread] 12:50:59.201 INFO [o.a.s.c.solrj.impl.HttpClientUtil ] Creating new http client, config:maxConnectionsPerHost=20maxConnections=1connTimeout=3socketTimeout=3retry=false [RecoveryThread] 12:50:59.377 INFO [org.apache.solr.update.PeerSync ] PeerSync: core=users url=url START replicas=[url] nUpdates=100 [RecoveryThread] 12:50:59.377 DEBUG [org.apache.solr.update.PeerSync ] PeerSync: core=users url=urlsolr startingVersions=0 [] [RecoveryThread] 12:50:59.390 WARN [org.apache.solr.update.PeerSync ] no frame of reference to tell of we've missed updates [RecoveryThread] 12:50:59.390 INFO [o.apache.solr.cloud.RecoveryStrategy] PeerSync Recovery was not successful - trying replication. core=users [RecoveryThread] 12:50:59.390 INFO [o.apache.solr.cloud.RecoveryStrategy] Starting Replication Recovery. core=users [RecoveryThread] 12:50:59.422 INFO [o.a.solr.common.cloud.ZkStateReader ] A cluster state change has occurred - updating... [localhost-startStop-1-EventThread] 12:50:59.575 INFO [o.a.s.c.solrj.impl.HttpClientUtil ] Creating new http client, config:maxConnections=128maxConnectionsPerHost=32followRedirects=false [RecoveryThread] 12:51:02.742 INFO [o.apache.solr.cloud.RecoveryStrategy] Begin buffering updates. core=users [RecoveryThread] 12:51:02.742 INFO [org.apache.solr.update.UpdateLog] Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null} [RecoveryThread] 12:51:02.743 INFO [o.apache.solr.cloud.RecoveryStrategy] Attempting to replicate from url. core=users [RecoveryThread] 12:51:02.743 INFO [o.a.s.c.solrj.impl.HttpClientUtil ] Creating new http client, config:maxConnections=128maxConnectionsPerHost=32followRedirects=false [RecoveryThread] 12:51:02.762 INFO [o.a.s.c.solrj.impl.HttpClientUtil ] Creating new http client, config:connTimeout=5000socketTimeout=2allowCompression=falsemaxConnections=1maxConnectionsPerHost=1 [RecoveryThread] 12:51:02.774 INFO [org.apache.solr.handler.SnapPuller ] No value set for 'pollInterval'. Timer Task not started. [RecoveryThread] 12:51:02.781 INFO [org.apache.solr.core.SolrCore ] SolrDeletionPolicy.onInit: commits:num=1 commit{dir=NRTCachingDirectory(org.apache.lucene.store.MMapDirectory@/solr/users/data/index lockFactory=org.apache.lucene.store.NativeFSLockFactory@6e28575; maxCacheMB=48.0 maxMergeSizeMB=4.0),segFN=segments_1,generation=1,filenames=[segments_1] [RecoveryThread] 12:51:02.782 INFO [org.apache.solr.core.SolrCore ] newest commit = 1 [RecoveryThread] 12:51:02.782 DEBUG [o.apache.solr.update.SolrIndexWriter] Opened Writer DirectUpdateHandler2 [RecoveryThread] 12:51:02.784 INFO [org.apache.solr.update.UpdateHandler] start commit{flags=0,_version_=0,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false} [RecoveryThread] 12:51:02.785 DEBUG [org.apache.solr.update.UpdateLog] TLOG: preCommit [RecoveryThread] 12:51:02.823 INFO [org.apache.solr.core.SolrCore ] SolrDeletionPolicy.onCommit: commits:num=2 commit{dir=NRTCachingDirectory(org.apache.lucene.store.MMapDirectory@/solr/users/data/index lockFactory=org.apache.lucene.store.NativeFSLockFactory@6e28575; maxCacheMB=48.0 maxMergeSizeMB=4.0),segFN=segments_1,generation=1,filenames=[segments_1] commit{dir=NRTCachingDirectory(org.apache.lucene.store.MMapDirectory@/solr/users/data/index lockFactory=org.apache.lucene.store.NativeFSLockFactory@6e28575; maxCacheMB=48.0 maxMergeSizeMB=4.0),segFN=segments_2,generation=2,filenames=[segments_2] [RecoveryThread] 12:51:02.824 INFO [org.apache.solr.core.SolrCore ] newest commit = 2 [RecoveryThread] 12:51:02.828 INFO [o.a.solr.search.SolrIndexSearcher ] Opening Searcher@5947fe65 main [RecoveryThread] 12:51:02.837 DEBUG [org.apache.solr.update.UpdateLog] TLOG: postCommit [RecoveryThread] 12:51:02.837 INFO
Need another way to boost relavence of recent content
I implemented the date boosting function outline here, placed into the boost parameter in the request handler: http://wiki.apache.org/solr/FunctionQuery#Date_Boosting Today it was reported to me that this boosting is producing terrible results. A close look at the description reveals that this isn't so much a boost on new content as it is a negative boost on old content. This isn't what I want. What I need to have happen is that content added today gets a small (and easily configurable) boost, content added yesterday gets a slightly smaller boost, tapering down to approximately two to four weeks in the past, at which point there would be no boost at all. A document that's a month old would have the same boost (none) as a document that's 50 years old. Can someone figure out the formula for this? Thanks, Shawn
RE: Need another way to boost relavence of recent content
Hi, Depending on use case the functions max, min, scale and map can be used really well to regulate the output of recip. Check their docs and you'll surely work it out. Perhaps scale will work best for you. Cheers -Original message- From:Shawn Heisey s...@elyograg.org Sent: Mon 14-Jan-2013 22:50 To: solr-user@lucene.apache.org Subject: Need another way to boost relavence of recent content I implemented the date boosting function outline here, placed into the boost parameter in the request handler: http://wiki.apache.org/solr/FunctionQuery#Date_Boosting Today it was reported to me that this boosting is producing terrible results. A close look at the description reveals that this isn't so much a boost on new content as it is a negative boost on old content. This isn't what I want. What I need to have happen is that content added today gets a small (and easily configurable) boost, content added yesterday gets a slightly smaller boost, tapering down to approximately two to four weeks in the past, at which point there would be no boost at all. A document that's a month old would have the same boost (none) as a document that's 50 years old. Can someone figure out the formula for this? Thanks, Shawn
RE: Need another way to boost relavence of recent content
Hi, Depending on use case the functions max, min, scale and map can be used really well to regulate the output of recip. Check their docs and you'll surely work it out. Perhaps scale will work best for you. Cheers -Original message- From:Shawn Heisey s...@elyograg.org Sent: Mon 14-Jan-2013 22:50 To: solr-user@lucene.apache.org Subject: Need another way to boost relavence of recent content I implemented the date boosting function outline here, placed into the boost parameter in the request handler: http://wiki.apache.org/solr/FunctionQuery#Date_Boosting Today it was reported to me that this boosting is producing terrible results. A close look at the description reveals that this isn't so much a boost on new content as it is a negative boost on old content. This isn't what I want. What I need to have happen is that content added today gets a small (and easily configurable) boost, content added yesterday gets a slightly smaller boost, tapering down to approximately two to four weeks in the past, at which point there would be no boost at all. A document that's a month old would have the same boost (none) as a document that's 50 years old. Can someone figure out the formula for this? Thanks, Shawn
I/O exception (java.net.SocketException) caught when processing request: Connection reset
I have a multi-threaded application in solrj 4. The threads (max 25) share one connection to HttpSolrServer. Each thread is running one query. This worked fine for a while, until it finally crashed with the following messages: Jan 12, 2013 12:52:15 PM org.apache.http.impl.client.DefaultRequestDirector tryExecute INFO: Retrying request Jan 12, 2013 12:52:15 PM org.apache.http.impl.client.DefaultRequestDirector tryExecute INFO: I/O exception (java.net.SocketException) caught when processing request: Connection reset I would like to catch this exception and reset the connection to the server. I don't get the whole stack trace with the above message, so I'm not sure where to do this. The only place I reference the server is for making a query with: server.query( query ) But the only exception this throws is SolrServerException, which I'm currently handling. Any suggestions would be greatly appreciated. FYI, this is how I'm setting up the initial server connection: server = new HttpSolrServer(url); server.setSoTimeout(0); server.setDefaultMaxConnectionsPerHost(50); server.setMaxTotalConnections(128); -- View this message in context: http://lucene.472066.n3.nabble.com/I-O-exception-java-net-SocketException-caught-when-processing-request-Connection-reset-tp4033309.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Need another way to boost relavence of recent content
On 1/14/2013 2:56 PM, Markus Jelsma wrote: Depending on use case the functions max, min, scale and map can be used really well to regulate the output of recip. Check their docs and you'll surely work it out. Perhaps scale will work best for you. I need someone to sanity check my work here. Here's my existing boost: str name=boostrecip(ms(NOW/DAY,pd),3.16e-11,1,1)/str After a careful look at your advice and the functions available, this is what I have come up with: min(recip(abs(ms(NOW/HOUR,pd)),3.85e-10,1.25,1),0.625) To get the second value for the recip function, I figured out how many milliseconds were in 30 days, then inverted that. If I understand everything correctly, this will result in boost values from 1.25 for docs created right now to 0.625 for docs created = 30 days ago. If I need to adjust 1.25 to X, then I also need to adjust the 0.625 value to 0.5 times X. Does that look OK? Thanks, Shawn
Re: RSS tutorial that comes with the apache-solr not indexing
Yes, thanks Alex, I've fixed solr/example/example-DIH/solr/rss/conf/rss-data-config.xml On Jan 14, 2013, at 2:08 PM, Alexandre Rafalovitch arafa...@gmail.com wrote: Is that something that needs to be updated in the example schema as well then? Regards, Alex. Personal blog: http://blog.outerthoughts.com/ LinkedIn: http://www.linkedin.com/in/alexandrerafalovitch - Time is the quality of nature that keeps events from happening all at once. Lately, it doesn't seem to be working. (Anonymous - via GTD book) On Mon, Jan 14, 2013 at 1:51 PM, bibhor bib...@gmail.com wrote: Hi Steve Thank you for your help. After I updated the rss-data-config to following, it worked. dataConfig dataSource type=URLDataSource / document entity name=rsstest pk=link url=http://www.feedforall.com/sample.xml; processor=XPathEntityProcessor forEach=/rss/channel/item transformer=DateFormatTransformer field column=title xpath=/rss/channel/item/title / field column=link xpath=/rss/channel/item/link / /entity /document /dataConfig -- View this message in context: http://lucene.472066.n3.nabble.com/RSS-tutorial-that-comes-with-the-apache-solr-not-indexing-tp4033067p4033254.html Sent from the Solr - User mailing list archive at Nabble.com.
Is there any way to check what index-time document boost value is?
Hello, I have indexed a document with an assigned document-level boost factor. Is there any way to double-check that the boost factor is actually recorded/used? I tried debug.explain.structured but it does not seem to have it. Regards, Alex. Personal blog: http://blog.outerthoughts.com/ LinkedIn: http://www.linkedin.com/in/alexandrerafalovitch - Time is the quality of nature that keeps events from happening all at once. Lately, it doesn't seem to be working. (Anonymous - via GTD book)
Re: incorrect solr update behavior
Of course, as soon as I post this, I discover this: https://issues.apache.org/jira/browse/SOLR-4134?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13537900#comment-13538174 i'll give this patch a spin in the morning. (this is not an example of how to use antecedents :)) -g On Mon, Jan 14, 2013 at 6:27 PM, Gary Yngve gary.yn...@gmail.com wrote: Posting this ?xml version=1.0 encoding=UTF-8?adddocfield name=nickname_s update=setblah/fieldfield name=tags_ss update=addqux/fieldfield name=tags_ss update=addquux/fieldfield name=idfoo/field/doc/add to an existing doc with foo and bar tags results in tags_ss containing arr name=tags_ss str{add=qux}/str str{add=quux}/str /arr whereas posting this ?xml version=1.0 encoding=UTF-8?adddocfield name=nickname_s update=setblah/fieldfield name=tags_ss update=addqux/fieldfield name=idfoo/field/doc/add results in the expected behavior: arr name=tags_ss strfoo/str strbar/str strqux/str /arr Any ideas? Thanks, Gary
RE: SOlr 3.5 and sharding
Ok that was my first thought... thanks for the confirmation -Original Message- From: Michael Ryan [mailto:mr...@moreover.com] Sent: January-14-13 10:06 AM To: solr-user@lucene.apache.org Subject: RE: SOlr 3.5 and sharding If you have the same documents -- with the same uniqueKey -- across multiple shards, the count will not be what you expect. You'll need to ensure that each document exists only on a single shard. -Michael -Original Message- From: Jean-Sebastien Vachon [mailto:jean-sebastien.vac...@wantedanalytics.com] Sent: Monday, January 14, 2013 9:59 AM To: solr-user@lucene.apache.org Subject: SOlr 3.5 and sharding Hi, I`m setting up a small Sorl setup consisting of 1 master node and 4 shards. For now, all four shards contains the exact same data. When I perform a query on each individual shards for the word `java` I am receiving the same number of docs (as expected). However, when I am going through the master node using the shards parameters, the number of results is slightly off by a few documents. There is nothing special in my setup so I`m looking for hints on why I am getting this problem Thanks - Aucun virus trouvé dans ce message. Analyse effectuée par AVG - www.avg.fr Version: 2013.0.2805 / Base de données virale: 2637/5996 - Date: 29/12/2012 La Base de données des virus a expiré.
Re: I/O exception (java.net.SocketException) caught when processing request: Connection reset
Hi, I suspect you might find some information about the cause on the server side, in your container's logs. I'd look there for the real source of the problem before trying to just reconnect from the client side. Otis -- Solr ElasticSearch Support http://sematext.com/ On Mon, Jan 14, 2013 at 5:06 PM, Joe joe.pol...@gmail.com wrote: I have a multi-threaded application in solrj 4. The threads (max 25) share one connection to HttpSolrServer. Each thread is running one query. This worked fine for a while, until it finally crashed with the following messages: Jan 12, 2013 12:52:15 PM org.apache.http.impl.client.DefaultRequestDirector tryExecute INFO: Retrying request Jan 12, 2013 12:52:15 PM org.apache.http.impl.client.DefaultRequestDirector tryExecute INFO: I/O exception (java.net.SocketException) caught when processing request: Connection reset I would like to catch this exception and reset the connection to the server. I don't get the whole stack trace with the above message, so I'm not sure where to do this. The only place I reference the server is for making a query with: server.query( query ) But the only exception this throws is SolrServerException, which I'm currently handling. Any suggestions would be greatly appreciated. FYI, this is how I'm setting up the initial server connection: server = new HttpSolrServer(url); server.setSoTimeout(0); server.setDefaultMaxConnectionsPerHost(50); server.setMaxTotalConnections(128); -- View this message in context: http://lucene.472066.n3.nabble.com/I-O-exception-java-net-SocketException-caught-when-processing-request-Connection-reset-tp4033309.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: performing a boolean query (OR) with a large number of terms
Hi, Also have a look at mm=0% acting as OR if you end up using dismax: http://wiki.apache.org/solr/ExtendedDisMax#mm_.28Minimum_.27Should.27_Match.29 Otis -- Solr ElasticSearch Support http://sematext.com/ On Thu, Jan 10, 2013 at 7:47 AM, Erick Erickson erickerick...@gmail.comwrote: No, you're pretty much on track. You can also just include the field multiple times if you want, itemModelNoExactMatchStr:123-4567 OR itemMOdelNoExactMatchStr:345-034985 But itemModelNoExactMatchStr:(123-4567 OR 345-034985) works just as well and is more compact. 15 terms is actually quite short by Solr standards, the default cap is 1024 boolean clauses, which you can change, it's just there to catch pathalogical (usually machine-generated) huge queries and make sure you consciously make a decision that it's OK. Best Erick On Wed, Jan 9, 2013 at 4:58 PM, geeky2 gee...@hotmail.com wrote: hello, environment: solr 3.5 i have a requirement to perform a boolean query (like the example below) with a large number of terms. the number of terms could be 15 or possibly larger. after looking over several theads and the smiley book - i think i just have include the parens and string all of the terms together with OR's i just want to make sure that i am not missing anything. is there a better or more efficient way of doing this? http://server:port /dir/core1/select?qt=modelItemNoSearchq=itemModelNoExactMatchStr:%285-100-NGRT7%20OR%205-10-10MS7%20OR%20404%29rows=30debugQuery=onrows=40 thx mark -- View this message in context: http://lucene.472066.n3.nabble.com/performing-a-boolean-query-OR-with-a-large-number-of-terms-tp4032039.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Is there any way to check what index-time document boost value is?
The norm function query gives you the combination of the index-time boost and length-normalization. And it's a low-resolution approximation at that. That's all that is stored, so that's as good as you can get. See: http://wiki.apache.org/solr/FunctionQuery#norm I believe that norm should show up in the debug explain. And for a description of the calculation of that norm factor, see: http://lucene.apache.org/core/4_0_0/core/org/apache/lucene/search/similarities/TFIDFSimilarity.html -- Jack Krupansky -Original Message- From: Alexandre Rafalovitch Sent: Monday, January 14, 2013 9:13 PM To: solr-user@lucene.apache.org Subject: Is there any way to check what index-time document boost value is? Hello, I have indexed a document with an assigned document-level boost factor. Is there any way to double-check that the boost factor is actually recorded/used? I tried debug.explain.structured but it does not seem to have it. Regards, Alex. Personal blog: http://blog.outerthoughts.com/ LinkedIn: http://www.linkedin.com/in/alexandrerafalovitch - Time is the quality of nature that keeps events from happening all at once. Lately, it doesn't seem to be working. (Anonymous - via GTD book)
Re: core.SolrCore - java.io.FileNotFoundException
The problem is occured recently again, this time exception is 2013-01-14 10:17:23,865 ERROR core.SolrCore - java.io.FileNotFoundException: /home/admin/index/core_p_shard4/index/_1ozb.fnm (No such file or directory) at java.io.RandomAccessFile.open(Native Method) at java.io.RandomAccessFile.init(RandomAccessFile.java:216) at org.apache.lucene.store.MMapDirectory.openInput(MMapDirectory.java:222) at org.apache.lucene.store.NRTCachingDirectory.openInput(NRTCachingDirectory.java:232) at org.apache.lucene.codecs.lucene40.Lucene40FieldInfosReader.read(Lucene40FieldInfosReader.java:52) at org.apache.lucene.index.SegmentCoreReaders.init(SegmentCoreReaders.java:101) at org.apache.lucene.index.SegmentReader.init(SegmentReader.java:57) at org.apache.lucene.index.ReadersAndLiveDocs.getReader(ReadersAndLiveDocs.java:120) at org.apache.lucene.index.BufferedDeletesStream.applyDeletes(BufferedDeletesStream.java:214) at org.apache.lucene.index.IndexWriter.applyAllDeletes(IndexWriter.java:3010) at org.apache.lucene.index.DocumentsWriter.applyAllDeletes(DocumentsWriter.java:180) at org.apache.lucene.index.DocumentsWriter.postUpdate(DocumentsWriter.java:310) at org.apache.lucene.index.DocumentsWriter.updateDocument(DocumentsWriter.java:386) at org.apache.lucene.index.IndexWriter.updateDocument(IndexWriter.java:1445) at org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandler2.java:210) at org.apache.solr.update.processor.RunUpdateProcessor.processAdd(RunUpdateProcessorFactory.java:61) at org.apache.solr.update.processor.UpdateRequestProcessor.processAdd(UpdateRequestProcessor.java:51) at org.apache.solr.update.processor.DistributedUpdateProcessor.versionAdd(DistributedUpdateProcessor.java:448) at org.apache.solr.update.processor.DistributedUpdateProcessor.processAdd(DistributedUpdateProcessor.java:325) at org.apache.solr.handler.loader.XMLLoader.processUpdate(XMLLoader.java:230) at org.apache.solr.handler.loader.XMLLoader.load(XMLLoader.java:157) But with INFOSTREAM open, I got more logs. here is some part of log which I think is valuable for this problem ordered by time. 1475075 IW 1 [Sun Jan 13 13:04:04 PST 2013; http-0.0.0.0-8080-11]: commit: pendingCommit != null 1475076 IW 1 [Sun Jan 13 13:04:04 PST 2013; http-0.0.0.0-8080-11]: commit: wrote segments file segments_1o 1475077 IFD 1 [Sun Jan 13 13:04:04 PST 2013; http-0.0.0.0-8080-11]: now checkpoint _320(4.0.0.2):C10 _32e(4.0.0.2):C89194 _32n(4.0.0.2):C9 8196 _333(4.0.0.2):C96199 _32m(4.0.0.2):C1 _32c(4.0.0.2):C1 _32q(4.0.0.2):C1 _32t(4.0.0.2):C1 _326(4.0.0.2):C53 _32k(4. 0.0.2):C8448/1 _332(4.0.0.2):C1 _335(4.0.0.2):C1 _336(4.0.0.2):C5799 _32r(4.0.0.2):C2477 _32v(4.0.0.2):C9042 _32x(4.0.0.2):C43 _334(4.0.0.2):C2328 _32z(4.0.0.2):C4035 _32o(4.0.0.2):C7201 _32s(4.0.0.2):C8020 [20 segments ; isCommit = true] 1475078 IFD 1 [Sun Jan 13 13:04:04 PST 2013; http-0.0.0.0-8080-11]: deleteCommits: now decRef commit segments_1n . 1490745 DWPT 0 [Mon Jan 14 09:21:37 PST 2013; http-0.0.0.0-8080-17]: flush postings as segment _1ozb numDocs=1 . 1491090 IFD 0 [Mon Jan 14 09:21:37 PST 2013; http-0.0.0.0-8080-4]: refresh [prefix=null]: removing newly created unreferenced file _1ozb.fdx 1491091 IFD 0 [Mon Jan 14 09:21:37 PST 2013; http-0.0.0.0-8080-4]: delete _1ozb.fdx . 1491152 IFD 0 [Mon Jan 14 09:21:37 PST 2013; http-0.0.0.0-8080-4]: refresh [prefix=null]: removing newly created unreferenced file _1ozb.fnm 1491153 IFD 0 [Mon Jan 14 09:21:37 PST 2013; http-0.0.0.0-8080-4]: delete _1ozb.fnm . 1491301 DW 0 [Mon Jan 14 09:21:37 PST 2013; http-0.0.0.0-8080-16]: publishFlushedSegment seg-private deletes=null 1491302 IW 0 [Mon Jan 14 09:21:37 PST 2013; http-0.0.0.0-8080-16]: publishFlushedSegment 1491303 BD 0 [Mon Jan 14 09:21:37 PST 2013; http-0.0.0.0-8080-16]: push deletes 10247 deleted terms (unique count=10247) bytesUsed=49152 delGe n=15715 packetCount=4599 totBytesUsed=117810176 1491304 IW 0 [Mon Jan 14 09:21:37 PST 2013; http-0.0.0.0-8080-16]: publish sets newSegment delGen=15716 seg=_1ozb(4.0.0.2):C1 1491305 IFD 0 [Mon Jan 14 09:21:37 PST 2013; http-0.0.0.0-8080-16]: now checkpoint _1oyu(4.0.0.2):C8992 _1oyv(4.0.0.2):C2030 _1oyz(4.0.0.2):C8 294 _1oyx(4.0.0.2):C2031 _1oyw(4.0.0.2):C259 _1oz3(4.0.0.2):C8375 _1oz1(4.0.0.2):C2836 _1oz5(4.0.0.2):C8231 _1oyy(4.0.0.2):C29 _1oz4(4. 0.0.2):C2988 _1oz8(4.0.0.2):C1 _1ozb(4.0.0.2):C1 [12 segments ; isCommit = false] 1491306 DW 0 [Mon Jan 14 09:21:37 PST 2013; http-0.0.0.0-8080-16]: force apply deletes bytesUsed=118015987 vs ramBuffer=1.34217728E8 149 1491307 BD 0 [Mon Jan 14 09:21:37 PST 2013; http-0.0.0.0-8080-2]: applyDeletes: infos=[_1oyu(4.0.0.2):C8992, _1oyv(4.0.0.2):C2030, _1oyz(4.0.0.2):C8294, _1oyx(4.0.0.2):C2031, _1oyw(4.0.0.2):C259, _1oz3(4.0.0.2):C8375,
Regarding Copyfield
hi in copy field i am not storing first_name last_name etc. but in dest = text it is showing first_name .. etc. in auto suggestion mode. my copy field are .. copyField source=percentage dest=text/ copyField source=university_name dest=text/ copyField source=course_name dest=text/ ... and field are .. field name=id type=text_general indexed=true stored=true required=true multiValued=false / field name=first_name type=text_general indexed=false stored=true/ field name=last_name type=text_general indexed=false stored=true/ field name=date_of_birth type=text_general indexed=false stored=true/ field name=state_name type=text_general indexed=false stored=true/ field name=mobile_no type=text_general indexed=false stored=true/ ... and also i want to make own field like text named as autosuggest then it is also not working for autosuggestion. please reply urgent -- View this message in context: http://lucene.472066.n3.nabble.com/Regarding-Copyfield-tp4033385.html Sent from the Solr - User mailing list archive at Nabble.com.
problem in Velocity spell output
if i am giving input antrag then it is showing following line. Did you mean {collationQuery=anurag,hits=1,misspellingsAndCorrections={aturag=anurag}}? i want output in format of Did you mean anurag ? how can i solve it please give me solution.. Thanks in advance Reply please urgent -- View this message in context: http://lucene.472066.n3.nabble.com/problem-in-Velocity-spell-output-tp4033390.html Sent from the Solr - User mailing list archive at Nabble.com.