Re: exceeded limit of maxWarmingSearchers ERROR
Nagendra You wrote, Naveen: *NRT with Apache Solr 3.3 and RankingAlgorithm does need a commit for a document to become searchable*. Any document that you add through update becomes immediately searchable. So no need to commit from within your update client code. Since there is no commit, the cache does not have to be cleared or the old searchers closed or new searchers opened, and warmed (error that you are facing). Looking at the link which you mentioned is clearly what we wanted. But the real thing is that you have RA does need a commit for a document to become searchable (please take a look at bold sentence) . In future, for more loads, can it cater to Master Slave (Replication) and etc to scale and perform better? If yes, we would like to go for NRT and looking at the performance described in the article is acceptable. We were expecting the same real time performance for a single user. What about multiple users, should we wait for 1-2 secs before calling the curl request to make SOLR perform better. Or internally it will handle with multiple request (multithreaded and etc). What would be doc size (10,000 docs) to allow JVM perform better? Have you done any kind of benchmarking in terms of multi threaded and multi user for NRT and also JVM tuning in terms of SOLR sever performance. Any kind of performance analysis would help us to decide quickly to switch over to NRT. Questions in terms for switching over to NRT, 1.Should we upgrade to SOLR 4.x ? 2. Any benchmarking (10,000 docs/secs). The question here is more specific the detail of individual doc (fields, number of fields, fields size, parameters affecting performance with faceting or w/o faceting) 3. What about multiple users ? A user in real time might be having an large doc size of .1 million. How to break and analyze which one is better (though it is our task to do). But still any kind of break up will help us. Imagine a user inbox. 4. JVM tuning and performance result based on Multithreaded environment. 5. Machine Details (RAM, CPU, and settings from SOLR perspective). Hoping that you are getting my point. We want to benchmark the performance. If you can involve me in your group, that would be great. Thanks Naveen 2011/8/15 Nagendra Nagarajayya nnagaraja...@transaxtions.com Bill: I did look at Marks performance tests. Looks very interesting. Here is the Apacle Solr 3.3 with RankingAlgorithm NRT performance: http://solr-ra.tgels.com/wiki/**en/Near_Real_Time_Search_ver_**3.xhttp://solr-ra.tgels.com/wiki/en/Near_Real_Time_Search_ver_3.x Regards - Nagendra Nagarajayya http://solr-ra.tgels.org http://rankingalgorithm.tgels.**org http://rankingalgorithm.tgels.org On 8/14/2011 7:47 PM, Bill Bell wrote: I understand. Have you looked at Mark's patch? From his performance tests, it looks pretty good. When would RA work better? Bill On 8/14/11 8:40 PM, Nagendra Nagarajayyannagarajayya@** transaxtions.com nnagaraja...@transaxtions.com wrote: Bill: The technical details of the NRT implementation in Apache Solr with RankingAlgorithm (SOLR-RA) is available here: http://solr-ra.tgels.com/**papers/NRT_Solr_**RankingAlgorithm.pdfhttp://solr-ra.tgels.com/papers/NRT_Solr_RankingAlgorithm.pdf (Some changes for Solr 3.x, but for most it is as above) Regarding support for 4.0 trunk, should happen sometime soon. Regards - Nagendra Nagarajayya http://solr-ra.tgels.org http://rankingalgorithm.tgels.**org http://rankingalgorithm.tgels.org On 8/14/2011 7:11 PM, Bill Bell wrote: OK, I'll ask the elephant in the roomŠ. What is the difference between the new UpdateHandler from Mark and the SOLR-RA? The UpdateHandler works with 4.0 does SOLR-RA work with 4.0 trunk? Pros/Cons? On 8/14/11 8:10 PM, Nagendra Nagarajayyannagarajayya@**transaxtions.comnnagaraja...@transaxtions.com wrote: Naveen: NRT with Apache Solr 3.3 and RankingAlgorithm does need a commit for a document to become searchable. Any document that you add through update becomes immediately searchable. So no need to commit from within your update client code. Since there is no commit, the cache does not have to be cleared or the old searchers closed or new searchers opened, and warmed (error that you are facing). Regards - Nagendra Nagarajayya http://solr-ra.tgels.org http://rankingalgorithm.tgels.**orghttp://rankingalgorithm.tgels.org On 8/14/2011 10:37 AM, Naveen Gupta wrote: Hi Mark/Erick/Nagendra, I was not very confident about NRT at that point of time, when we started project almost 1 year ago, definitely i would try NRT and see the performance. The current requirement was working fine till we were using commitWithin 10 millisecs in the XMLDocument which we were posting to SOLR. But due to which, we were getting very poor performance (almost 3 mins for 15,000 docs) per user. There are many paraller user committing to our SOLR. So we removed the commitWithin, and hence
Re: how to add automatic keyword filters in solr?
ex: if i type letter a i should be able to see the related words in dropdown manner like ant,apple,ape,axe..etc... One of the many other solutions : http://wiki.apache.org/solr/Suggester
OOM due to JRE Issue (LUCENE-1566)
Hi, This might probably have been discussed long time back, but I got this error recently in one of my production slaves. SEVERE: java.lang.OutOfMemoryError: OutOfMemoryError likely caused by the Sun VM Bug described in https://issues.apache.org/jira/browse/LUCENE-1566; try calling FSDirectory.setReadChunkSize with a a value smaller than the current chunk size (2147483647) I am currently using Solr1.4. Going through JIRA Issue comments, I found that this patch applies to 2.9 or above. We are also planning an upgrade to Solr 3.3. Is this patch included in 3.3 so as to I don't have to manually apply the patch? What are the other workarounds of the problem? Thanks in adv. *Pranav Prakash* temet nosce Twitter http://twitter.com/pranavprakash | Blog http://blog.myblive.com | Google http://www.google.com/profiles/pranny
Filtering results based on a set of values for a field
Hello, Solrs we are trying to filter out documents written by (one or more of) the authors from a mediumish list (~2K). The document set itself is in the millions. Apart from the obvious approach of building a huge OR-list and appending it to the query, it seems that writing a Lucene[1] filter (or a SolrFilter[2]) seems to suggest itself. In fact [3] seems to strongly encourage this approach. Basically, as we understand it, the filter's method getDocIdSet gets called and is fed with index segments, one spoonful at a time. It then decides which docs of the segment will be accepted, setting the corresponding bits in the result (in our case, e.g. look up the document's author's name in a HashMap or something like it). Our first question is: how does it all fit together? Would be enough to write such a class? How do I reference that in the SOLR configuration? In the query? A Lucene Filter or a SolrFilter? The problem is, we are experiencing very slow response times, in the order of 12 seconds for a query (the OR alternative, which we tested on a smallish author list of aboug a couple of hundred is nearly-instantaneous). Our second question is: are we on track with this? Intuition would say, of course, that sifting sequentially through the index, checking each document for its author *will* take its time. So may be the approach is doomed? Are there other, better approaches? Thanks for any pointers -- [1] https://builds.apache.org/job/Lucene-3.x/javadoc/all/org/apache/lucene/search/Filter.html?is-external=true [2] http://lucene.apache.org/solr/api/org/apache/solr/search/SolrFilter.html [1] http://wiki.apache.org/lucene-java/FilteringOptions -- tomás
Re: how to add automatic keyword filters in solr?
Thank u very much for ur reply iorixxx i already know about the field type suggest and i am able to get the related keywords in json response format my problem is i developed one jsp page and integrated to solr if i type test in the jsp page i will get the response whichever having the test keyword...everything is fine ...but,when ever i type t i need to get the full keywords like test,Test,TEST etc...in the jsp page search field ..i am unable to do this so,please help me to get this in jsp page searchfield.. -- View this message in context: http://lucene.472066.n3.nabble.com/how-to-add-automatic-keyword-filters-in-solr-tp3257708p3258058.html Sent from the Solr - User mailing list archive at Nabble.com.
how to use solr search keywords url in jsp page?
hi everybody i developed one response format which gives the following response when ever i enter t in my solr search field... { responseHeader:{ status:0, QTime:0, params:{ fl:keywords, indent:on, start:0, q:t\r\n, version:2.2, rows:30}}, response:{numFound:19,start:0,docs:[ { keywords:test}, { keywords:test}, { keywords:test}, { keywords:Test}, { keywords:test}, { keywords:test}, { keywords:test}, { keywords:test}, { keywords:test}, { keywords:Test}, { keywords:Test}, { keywords:Test}, { keywords:Test}, { keywords:Test}, { keywords:Test}, { keywords:test}, { keywords:testing}, { keywords:testing}, { keywords:test iphone android}] }} now i dont know how to implement this in my jsp pageactually i have one jsp page with search field when ever i enter the word t in my search field i need to get all the above results as auto keywords as like in google...if we enter a the related words ant,axeetc...like this i need to display please help me thank u in advance -- View this message in context: http://lucene.472066.n3.nabble.com/how-to-use-solr-search-keywords-url-in-jsp-page-tp3258100p3258100.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: OOM due to JRE Issue (LUCENE-1566)
hey, On Tue, Aug 16, 2011 at 9:34 AM, Pranav Prakash pra...@gmail.com wrote: Hi, This might probably have been discussed long time back, but I got this error recently in one of my production slaves. SEVERE: java.lang.OutOfMemoryError: OutOfMemoryError likely caused by the Sun VM Bug described in https://issues.apache.org/jira/browse/LUCENE-1566; try calling FSDirectory.setReadChunkSize with a a value smaller than the current chunk size (2147483647) I am currently using Solr1.4. Going through JIRA Issue comments, I found that this patch applies to 2.9 or above. We are also planning an upgrade to Solr 3.3. Is this patch included in 3.3 so as to I don't have to manually apply the patch? AFAIK, solr 1.4 is on Lucene 2.9.1 so this patch is already applied to the version you are using. maybe you can provide the stacktrace and more deatails about your problem and report back? simon What are the other workarounds of the problem? Thanks in adv. *Pranav Prakash* temet nosce Twitter http://twitter.com/pranavprakash | Blog http://blog.myblive.com | Google http://www.google.com/profiles/pranny
SOlr upgrade: Invalid version (expected 2, but 1) error when using shards
Hi, I just migrated to solr 3.3 from 1.4.1. My index is still in 1.4.1 format (will be migrated soon). I have an error when I use sharding with the new version: org.apache.solr.common.SolrException: java.lang.RuntimeException: Invalid version (expected 2, but 1) or the data in not in 'javabin' format However, if I request each shard independently (/request), answer is correct. So the error is triggered only with the shard mechanism. While I foresee to upgrade my indexes, I'd like to understand the issue, e.g. is it an upgrade issue or don't shards support using an old format. Thanks Olivier
Re: OOM due to JRE Issue (LUCENE-1566)
AFAIK, solr 1.4 is on Lucene 2.9.1 so this patch is already applied to the version you are using. maybe you can provide the stacktrace and more deatails about your problem and report back? Unfortunately, I have only this much information with me. However following is my speficiations, if they are any helpful :- /usr/bin/java -d64 -Xms5000M -Xmx5000M -XX:+UseParallelGC -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -Xloggc:$GC_LOGFILE -XX:+CMSPermGenSweepingEnabled -Dsolr.solr.home=multicore -Denable.slave=true -jar start.jar 32GiB RAM Any thoughts? Will a switch to ConcurrentGC help in any means?
How to deal with java.net.SocketTimeoutException: Read timed out on commit?
Hi, We're getting a lot of these timeouts during bulk feeding or a large document set. We're sending batches of 1000 documents and commiting every 15 minutes or for every 10.000 docs, whichever happens first. We find that the first few commits (after 10'/20'/30' docs) go through without exceptions (although the CommonsHttpSolrServer.soTimeout is set quite high - 12ms). The time for a commit to complete seems to increase for every subsequent commit request, and eventually the timeout value is reached causing Solrj to throw the java.net.SocketTimeoutException. Increasing the soTimeout value only delays the problem and it will allow us to send a few more commits before seeing the exception. I've noticed that even though we get these exceptions on commit, documents seem to be searchable after some time. Does that mean that the commit process in solr continues even though the socket times out? We're also setting waitFlush=false and waitSearcher=false when committing, which I would expect to enable a non-blocking commit request to solr, but on n+1 commits this doesn't aleays happen and this when we're getting the timeouts. I suspect that solr.commit(false,false) is a non-blocking call only if a all previous commits are complete. Correct? I'm aware that commiting often is not adviced as it is a resouce intensive process for solr, however it is a requirement that documents are searchable before the entire document set has been added to solr. Our workaround is to limit the amount of commits by time and document count triggers in our code. We cannot use the autocommit feature in solrconfig.xml because we also need to be able to issue immediate commits for incremental feeding. So, my main question is regarding the severity of java.net.SocketTimeoutException: Read timed out on commit. Should I worry and try to re-issue commit when this happens, or is it safe to ignore it? Thanks for any input! Cheers, Geir
Re: How to deal with java.net.SocketTimeoutException: Read timed out on commit?
We too were getting same issue. We solved it by ensuring that when commit is in progress, no one access the index. Though SOLR's UpdateRequest does it, we will still read timeout issues because of CommonsHttpSolrServer. If we have another layer which doesnt send the request itself, then you wont get this error. On Tue, Aug 16, 2011 at 5:34 PM, Geir Gullestad Pettersen gei...@gmail.comwrote: Hi, We're getting a lot of these timeouts during bulk feeding or a large document set. We're sending batches of 1000 documents and commiting every 15 minutes or for every 10.000 docs, whichever happens first. We find that the first few commits (after 10'/20'/30' docs) go through without exceptions (although the CommonsHttpSolrServer.soTimeout is set quite high - 12ms). The time for a commit to complete seems to increase for every subsequent commit request, and eventually the timeout value is reached causing Solrj to throw the java.net.SocketTimeoutException. Increasing the soTimeout value only delays the problem and it will allow us to send a few more commits before seeing the exception. I've noticed that even though we get these exceptions on commit, documents seem to be searchable after some time. Does that mean that the commit process in solr continues even though the socket times out? We're also setting waitFlush=false and waitSearcher=false when committing, which I would expect to enable a non-blocking commit request to solr, but on n+1 commits this doesn't aleays happen and this when we're getting the timeouts. I suspect that solr.commit(false,false) is a non-blocking call only if a all previous commits are complete. Correct? I'm aware that commiting often is not adviced as it is a resouce intensive process for solr, however it is a requirement that documents are searchable before the entire document set has been added to solr. Our workaround is to limit the amount of commits by time and document count triggers in our code. We cannot use the autocommit feature in solrconfig.xml because we also need to be able to issue immediate commits for incremental feeding. So, my main question is regarding the severity of java.net.SocketTimeoutException: Read timed out on commit. Should I worry and try to re-issue commit when this happens, or is it safe to ignore it? Thanks for any input! Cheers, Geir
RE: ideas for indexing large amount of pdf docs
Hi Jay, thanks. great idea, in next days we'll try to do something like you'd exposed. best, rode. --- Rode González Libnova, SL Paseo de la Castellana, 153-Madrid [t]91 449 08 94 [f]91 141 21 21 www.libnova.es -Mensaje original- De: Jaeger, Jay - DOT [mailto:jay.jae...@dot.wi.gov] Enviado el: lunes, 15 de agosto de 2011 14:54 Para: solr-user@lucene.apache.org Asunto: RE: ideas for indexing large amount of pdf docs Note on i: Solr replication provides pretty good clustering support out-of-the-box, including replication of multiple cores. Read the Wiki on replication (Google +solr +replication if you don't know where it is). In my experience, the problem with indexing PDFs is it takes a lot of CPU on the document parsing side (client), not on the Solr server side. So make sure you do that part on the client and not the server. Avoiding iii: Suggest that you write yourself a multi-threaded performance test so that you aren't guessing what your performance will be. We wrote one in Perl. It handles an individual thread (we were testing inquiry), and we wrote a little batch file / shell script to start up the desired number of threads. The main statement in our batch file (the rest just set the variables). A Shell script would be even easier. for /L %%i in (1,1,%THREADS%) DO start /B perl solrtest.pl -h %SOLRHOST% -c %COUNT% -u %1 -p %2 -r %SOLRREALM% -f %SOLRLOC%\firstsynonyms.txt -l %SOLRLOC%\lastsynonyms.txt -z %FUZZ% The perl #!/usr/bin/perl # # Perl program to run a thread of solr testing # use Getopt::Std; # For options processing use POSIX;# For time formatting use XML::Simple; # For processing of XML config file use Data::Dumper; # For debugging XML config file use HTTP::Request::Common;# For HTTP request to Solr use HTTP::Response; use LWP::UserAgent; # For HTTP request to Solr $host = YOURHOST:8983; $realm = YOUR AUTHENTICATION REALM; $firstlist = firstsynonyms.txt; $lastlist = lastsynonyms.txt; $fuzzy = ; $me = $0; sub usage() { print perl $me -c iterations [-d] [-h host:port ] [-u user [-p password]] \n; print \t\t[-f firstnamefile] [-l lastnamefile] [-z fuzzy] [-r realm]\n; exit(8); } # # Process the command line options, and open the output file. # getopts('dc:u:p:f:l:h:r:z:') || usage(); if(!$opt_c) { usage(); } $count = $opt_c; if($opt_u) { $user = $opt_u; } if($opt_p) { $password = $opt_p; } if($opt_h) { $host = $opt_h; } if($opt_f) { $firstlist = $opt_f; } if($opt_l) { $lastlist = $opt_l; } if($opt_r) { $realm = $opt_r; } if($opt_z) { $fuzzy = ~ . $opt_z; } $debug = $opt_d; # # If the host string does not include a :, add :80 # if($host !~ /:/) { $host = $host . :80; } # # Read the lists of first and last names # open(SYNFILE,$firstlist) || die Can't open first name list $firstlist\n; while(SYNFILE) { @newwords = split /,/; for($i=0; $i = $#newwords; ++$i) { $newwords[$i] =~ s/^\s+//; $newwords[$i] =~ s/\s+$//; $newwords[$i] = lc($newwords[$i]); } push @firstnames, @newwords; } close(SYNFILE); open(SYNFILE,$lastlist) || die Can't open last name list $lastlist\n; while(SYNFILE) { @newwords = split /,/; for($i=0; $i = $#newwords; ++$i) { $newwords[$i] =~ s/^\s+//; $newwords[$i] =~ s/\s+$//; $newwords[$i] = lc($newwords[$i]); } push @lastnames, @newwords; } close(SYNFILE); print $#firstnames First Names, $#lastnames Last Names\n; print User: $user\n; my $userAgent = LWP::UserAgent-new(agent = 'solrtest.pl'); $userAgent-credentials($host,$realm,$user,$password); $uri = http://$host/solr/select;; $starttime = time(); for($c=0; $c $count; ++$c) { $fname = $firstnames[rand $#firstnames]; $lname = $lastnames[rand $#lastnames]; $response = $userAgent-request( POST $uri, [ q = lnamesyn:$lname AND fnamesyn:$fname$fuzzy, rows = 25 ]); if($debug) { print Query: lnamesyn:$lname AND fnamesyn:$fname$fuzzy; print $response-content(); } print POST for $fname $lname completed, HTTP status= . $response-code . \n; } $elapsed = time() - $starttime; $average = $elapsed / $count; print Time: $elapsed s ($average/request)\n; -Original Message- From: Rode Gonzalez (libnova) [mailto:r...@libnova.es] Sent: Saturday, August 13, 2011 3:50 AM To: solr-user@lucene.apache.org Subject: ideas for indexing large amount of pdf docs Hi all, I want to ask about the best way to implement
RE: Product data schema question
On the surface, you could simply add some more fields to your schema. But as far as I can tell, you would have to have a separate Solr document for each SKU/size combination, and store the rest of the information (brand, model, color, SKU) redundantly and make the unique key a combination of the SKU and the size (presumably by having an additional field called sku_size - as far as I can tell Solr can't make a key up out of multiple fields). But, perhaps you shouldn't store that dynamic inventory information in Solr. Instead store a key that gets you to an inventory database organized by SKU and size. In such a schema, Size could be added as a multi-valued field to your existing schema, so you would know what sizes existed, so you could at least facet on that, maybe. That way Solr doesn't have to be updated every time the inventory changes. Of course, that won't help with the faceting on variant inventory. -Original Message- From: Steve Cerny [mailto:sjce...@gmail.com] Sent: Monday, August 15, 2011 6:29 PM To: solr-user@lucene.apache.org Subject: Product data schema question I'm working on an online eCommerce project and am having difficulties building the core / index schema. Here is the way we organize our product information in a normalized database. A product model has many SKUs (called colorways) A SKU has many sizes (called variants) A SKU size has associated inventory (called variant inventory) When we setup our product core we have the following field information Doc * brand * model name * SKU * color name Sample records are as follows * Haynes, Undershirt, 1234, white * Haynes, Undershirt, 1235, grey * Fruit of the Loom, Undershirt, 1236, white * Fruit of the Loom, Underwear, 1237, grey The issue I'm having is I want to add inventory to each size of each SKU for faceting. Example, SKU 1234 has sizes small, medium, large. Size small has 5 in stock, size medium 10, and size large 25. In a normalized data table I would have a separate table just for inventory and related it back to the SKU with a foreign key. How do I store size and inventory information effectively with Solr? -- Steve
Re: Can Master push data to slave
Why do you care about the lock file on the slave? It shouldn't matter, so I'm wondering if this is an XY problem: From Hossman's Apache page: Your question appears to be an XY Problem ... that is: you are dealing with X, you are assuming Y will help you, and you are asking about Y without giving more details about the X so that we can understand the full issue. Perhaps the best solution doesn't involve Y at all? See Also: http://www.perlmonks.org/index.pl?node_id=542341 Best Erick On Mon, Aug 15, 2011 at 1:59 AM, Pawan Darira pawan.dar...@gmail.com wrote: Regarding point b, i mean that when Slave server does a replication from Master, it creates a lock-file in it's index directory. How to avoid that? On Tue, Aug 9, 2011 at 2:56 AM, Markus Jelsma markus.jel...@openindex.iowrote: Hi, Hi I am using Solr 1.4. and doing a replication process where my slave is pulling data from Master. I have 2 questions a. Can Master push data to slave Not in current versions. Not sure about exotic patches for this. b. How to make sure that lock file is not created while replication What do you mean? Please help thanks Pawan
Re: Minimum score filter
Right, so you're using edismax? This is expected. You can do a number of things: 1 change the parameters of edismax 2 have your app filter out returns that dive beneath some threshold that is relative to the score of the first doc in the list. But I don't see why, given your example, it makes a difference. Because you got the document you were looking for in this instance says nothing abou the next time when you search on something different Best Erick On Mon, Aug 15, 2011 at 2:22 PM, Donald J. Organ IV dor...@donaldorgan.com wrote: OK I am doing a search using the following fields name^2.0 code^1.8 cat_search^1.5 description^0.8 I am searching for: free range dog nips I am getting back 2 documents the first is the document I am looking for, and contains those works in the name field, as the name field is Free Range Dog Nip Chicken Breast Wraps The second looks like its matching because those words are contained within the description. - Original Message - From: simon mtnes...@gmail.com To: solr-user@lucene.apache.org Sent: Monday, August 15, 2011 1:59:17 PM Subject: Re: Minimum score filter The absolute value of a relevance score doesn't have a lot of meaning and the range of scores can vary a lot depending on any boost you may apply. Even if you normalize them (say on a 1-100 scale where 100 is the max relevance) you can't really draw any valid conclusions from those values. It would help if you described exactly what problem you're trying to solve. -Simon On Mon, Aug 15, 2011 at 1:02 PM, Donald J. Organ IV
Re: Indexing from a database via SolrJ
What have you tried and what doesn't it do that you want it to do? This works, instantiating the StreamingUpdateSolrServer (server) and the JDBC connection/SQL statement are left as exercises for the reader G.: while (rs.next()) { SolrInputDocument doc = new SolrInputDocument(); String id = rs.getString(id); String title = rs.getString(title); String text = rs.getString(text); doc.addField(id, id); doc.addField(title, title); doc.addField(text, text); docs.add(doc); ++counter; ++total; if (counter 100) { // Completely arbitrary, just batch up more than one document for throughput! server.add(docs); docs.clear(); counter = 0; } } Best Erick On Mon, Aug 15, 2011 at 6:25 PM, Shawn Heisey s...@elyograg.org wrote: Is there a simple way to get all the fields from a jdbc resultset into a bunch of SolrJ documents, which I will then send to be indexed in Solr? I would like to avoid the looping required to copy the data one field at a time. Copying it one document at a time would be acceptable, but it would be nice if there was a way to copy them all at once. Another idea that occurred to me is to add the dataimporter jar to my project and leverage it to do the heavy lifting, but I will need some pointers about what objects and methods to research. Is that a reasonable idea, or is it too integrated into the server code to be used with SolrJ? Can anyone point me in the right direction? Thanks, Shawn
Re: Migration from Autonomy IDOL to SOLR
Hi Arcadius, currently we have a migration project from verity k2 search server to solr. I do not know IDOL, but autonomy bought verity before IDOL was released, so possible it is comparable? verity k2 works directly on xml-Files, in result the query syntax is a little bit like xpath e.g. with text1 IN zone2 IN zone1 instead of contains(//zone1/zone2,'text1'). About verity query syntax: http://gregconely.getmyip.com/dl/OTG%20Software/5.30.087%20Suite%20%28SP3%29/Disc%204%20-%20Verity/Verity%20K2%20Server%205.5/doc/docs/pdf/VerityQueryLanguage.pdf Does IDOL work the same way? Best regards Karsten P.S. in Context: http://lucene.472066.n3.nabble.com/Migration-from-Autonomy-IDOL-to-SOLR-td3255377.html Original-Nachricht Datum: Mon, 15 Aug 2011 11:11:36 +0100 Von: Arcadius Ahouansou arcad...@menelic.com An: solr-user@lucene.apache.org Betreff: Migration from Autonomy IDOL to SOLR Hello. We have a couple of application running on half a dozen Autonomy IDOL servers. Currently, all feature we need are supported by Solr. We have done some internal testing and realized that SOLR would do a better job. So, we are investigation all possibilities for a smooth migration from IDOL to SOLR. I am looking for advice from people who went through something similar. Ideally, we would like to keep most of our legacy code unchanged and have a kind of query-translation-layer plugged into our app if possible. -Is there lib available? -Any thought? Thanks. Arcadius.
Exact matching on names?
Hi all- I'm missing something fundamental yet I've been unable to find the definitive answer for exact name matching. I'm indexing names using the standard text field type and my search is for the name clarke. My results include clark, which is incorrect, it needs to match clarke exactly (case insensitive). I tried textType but that doesn't work because I believe it needs to be *really* exact, whereas I'm looking for things like clark oil, bob, frank, and clark, etc. Thanks for any help, Ron DISCLAIMER: This electronic message, including any attachments, files or documents, is intended only for the addressee and may contain CONFIDENTIAL, PROPRIETARY or LEGALLY PRIVILEGED information. If you are not the intended recipient, you are hereby notified that any use, disclosure, copying or distribution of this message or any of the information included in or with it is unauthorized and strictly prohibited. If you have received this message in error, please notify the sender immediately by reply e-mail and permanently delete and destroy this message and its attachments, along with any copies thereof. This message does not create any contractual obligation on behalf of the sender or Law Bulletin Publishing Company. Thank you.
Re: SOlr upgrade: Invalid version (expected 2, but 1) error when using shards
On 8/16/2011 4:16 AM, olivier sallou wrote: I just migrated to solr 3.3 from 1.4.1. My index is still in 1.4.1 format (will be migrated soon). I have an error when I use sharding with the new version: org.apache.solr.common.SolrException: java.lang.RuntimeException: Invalid version (expected 2, but 1) or the data in not in 'javabin' format However, if I request each shard independently (/request), answer is correct. So the error is triggered only with the shard mechanism. In version 3.1, the javabin format was changed, it's completely incompatible with 1.4.1 and earlier at the application level, which should have nothing to do with the index format. Javabin is used in distributed search and by SolrJ. It sounds like you either have an old SolrJ version talking to the new Solr version, or you are requesting data from your shards using an old version of Solr. If it's the latter, just upgrade the Solr instance that serves as the broker/middleman between your application and your shards. If it's SolrJ, you have two choices. You can either upgrade your SolrJ jar and the jars found in solrj-lib, or keep using the old version and add the following line to your code right after your server object is created. This tells SolrJ to use the XML format, which works on any version: server.setParser(new XMLResponseParser()); You'll also need to import org.apache.solr.client.solrj.impl.XMLResponseParser for the above line to work. Thanks, Shawn
Re: Indexing from a database via SolrJ
On 8/16/2011 7:14 AM, Erick Erickson wrote: What have you tried and what doesn't it do that you want it to do? This works, instantiating the StreamingUpdateSolrServer (server) and the JDBC connection/SQL statement are left as exercises for the readerG.: while (rs.next()) { SolrInputDocument doc = new SolrInputDocument(); String id = rs.getString(id); String title = rs.getString(title); String text = rs.getString(text); doc.addField(id, id); doc.addField(title, title); doc.addField(text, text); docs.add(doc); ++counter; ++total; if (counter 100) { // Completely arbitrary, just batch up more than one document for throughput! server.add(docs); docs.clear(); counter = 0; } } I've implemented a basic loop with the structure you've demonstrated, but it currently doesn't do anything yet with SolrInputDocument or SolrDocumentList. I figured there would be a way to avoid going through the field list one by one, but what you've written suggests that the field-by-field method is required. I can live with that. It does look like addField just takes an Object, so hopefully I can create a loop that determines the type of each field from the JDBC metadata, retrieves the correct Java type from the ResultSet, and inserts it. I imagine that everything still works if you happen to insert a field that doesn't exist in the index. This must be how the DIH does it, so I was hoping that the DIH might expose a method that takes a ResultSet and produces a SolrDocumentList. I still have to take a deeper look at the source and documentation. Thanks for the help so far, I can get a little more implemented now. Shawn
Re: Exact matching on names?
exact can mean a lot of things (do diacritics count?, etc), but in this case, it sounds like you just need to turn off the stemmer you have on this fieldtype (or create a new one that doesn't include the stemmer). hth, rob On Tue, Aug 16, 2011 at 11:20 AM, Olson, Ron rol...@lbpc.com wrote: Hi all- I'm missing something fundamental yet I've been unable to find the definitive answer for exact name matching. I'm indexing names using the standard text field type and my search is for the name clarke. My results include clark, which is incorrect, it needs to match clarke exactly (case insensitive). I tried textType but that doesn't work because I believe it needs to be *really* exact, whereas I'm looking for things like clark oil, bob, frank, and clark, etc. Thanks for any help, Ron DISCLAIMER: This electronic message, including any attachments, files or documents, is intended only for the addressee and may contain CONFIDENTIAL, PROPRIETARY or LEGALLY PRIVILEGED information. If you are not the intended recipient, you are hereby notified that any use, disclosure, copying or distribution of this message or any of the information included in or with it is unauthorized and strictly prohibited. If you have received this message in error, please notify the sender immediately by reply e-mail and permanently delete and destroy this message and its attachments, along with any copies thereof. This message does not create any contractual obligation on behalf of the sender or Law Bulletin Publishing Company. Thank you.
Re: Product data schema question
Jay, this is great information. I don't know enough about Solr whether this is possible...Can we setup two indexes in the same core, one for product_catalog and the other for inventory? Then using a Solr query we could join the indexed content together. In Sql it would look like this select p.brand , p.model_name , p.sku , p.color_name , i.variant_id , i.variant_count from product_catalog p join inventory i on (p.sku = i.sku) On Tue, Aug 16, 2011 at 8:00 AM, Jaeger, Jay - DOT jay.jae...@dot.wi.govwrote: On the surface, you could simply add some more fields to your schema. But as far as I can tell, you would have to have a separate Solr document for each SKU/size combination, and store the rest of the information (brand, model, color, SKU) redundantly and make the unique key a combination of the SKU and the size (presumably by having an additional field called sku_size - as far as I can tell Solr can't make a key up out of multiple fields). But, perhaps you shouldn't store that dynamic inventory information in Solr. Instead store a key that gets you to an inventory database organized by SKU and size. In such a schema, Size could be added as a multi-valued field to your existing schema, so you would know what sizes existed, so you could at least facet on that, maybe. That way Solr doesn't have to be updated every time the inventory changes. Of course, that won't help with the faceting on variant inventory. -Original Message- From: Steve Cerny [mailto:sjce...@gmail.com] Sent: Monday, August 15, 2011 6:29 PM To: solr-user@lucene.apache.org Subject: Product data schema question I'm working on an online eCommerce project and am having difficulties building the core / index schema. Here is the way we organize our product information in a normalized database. A product model has many SKUs (called colorways) A SKU has many sizes (called variants) A SKU size has associated inventory (called variant inventory) When we setup our product core we have the following field information Doc * brand * model name * SKU * color name Sample records are as follows * Haynes, Undershirt, 1234, white * Haynes, Undershirt, 1235, grey * Fruit of the Loom, Undershirt, 1236, white * Fruit of the Loom, Underwear, 1237, grey The issue I'm having is I want to add inventory to each size of each SKU for faceting. Example, SKU 1234 has sizes small, medium, large. Size small has 5 in stock, size medium 10, and size large 25. In a normalized data table I would have a separate table just for inventory and related it back to the SKU with a foreign key. How do I store size and inventory information effectively with Solr? -- Steve -- Steve
RE: Product data schema question
No, I don't think so. A given core can only use one configuration and therefore only one schema, as far as I know, and a schema can only have one key. You could use two cores with two configurations (but that presumably wouldn't be much help). Solr is not a DBMS. It is an index. -Original Message- From: Steve Cerny [mailto:sjce...@gmail.com] Sent: Tuesday, August 16, 2011 11:37 AM To: solr-user@lucene.apache.org Subject: Re: Product data schema question Jay, this is great information. I don't know enough about Solr whether this is possible...Can we setup two indexes in the same core, one for product_catalog and the other for inventory? Then using a Solr query we could join the indexed content together. In Sql it would look like this select p.brand , p.model_name , p.sku , p.color_name , i.variant_id , i.variant_count from product_catalog p join inventory i on (p.sku = i.sku) On Tue, Aug 16, 2011 at 8:00 AM, Jaeger, Jay - DOT jay.jae...@dot.wi.govwrote: On the surface, you could simply add some more fields to your schema. But as far as I can tell, you would have to have a separate Solr document for each SKU/size combination, and store the rest of the information (brand, model, color, SKU) redundantly and make the unique key a combination of the SKU and the size (presumably by having an additional field called sku_size - as far as I can tell Solr can't make a key up out of multiple fields). But, perhaps you shouldn't store that dynamic inventory information in Solr. Instead store a key that gets you to an inventory database organized by SKU and size. In such a schema, Size could be added as a multi-valued field to your existing schema, so you would know what sizes existed, so you could at least facet on that, maybe. That way Solr doesn't have to be updated every time the inventory changes. Of course, that won't help with the faceting on variant inventory. -Original Message- From: Steve Cerny [mailto:sjce...@gmail.com] Sent: Monday, August 15, 2011 6:29 PM To: solr-user@lucene.apache.org Subject: Product data schema question I'm working on an online eCommerce project and am having difficulties building the core / index schema. Here is the way we organize our product information in a normalized database. A product model has many SKUs (called colorways) A SKU has many sizes (called variants) A SKU size has associated inventory (called variant inventory) When we setup our product core we have the following field information Doc * brand * model name * SKU * color name Sample records are as follows * Haynes, Undershirt, 1234, white * Haynes, Undershirt, 1235, grey * Fruit of the Loom, Undershirt, 1236, white * Fruit of the Loom, Underwear, 1237, grey The issue I'm having is I want to add inventory to each size of each SKU for faceting. Example, SKU 1234 has sizes small, medium, large. Size small has 5 in stock, size medium 10, and size large 25. In a normalized data table I would have a separate table just for inventory and related it back to the SKU with a foreign key. How do I store size and inventory information effectively with Solr? -- Steve -- Steve
Re: Product data schema question
Thanks Jay, if we come to a reasonable solution are you interested in the details? On Tue, Aug 16, 2011 at 11:44 AM, Jaeger, Jay - DOT jay.jae...@dot.wi.govwrote: No, I don't think so. A given core can only use one configuration and therefore only one schema, as far as I know, and a schema can only have one key. You could use two cores with two configurations (but that presumably wouldn't be much help). Solr is not a DBMS. It is an index. -Original Message- From: Steve Cerny [mailto:sjce...@gmail.com] Sent: Tuesday, August 16, 2011 11:37 AM To: solr-user@lucene.apache.org Subject: Re: Product data schema question Jay, this is great information. I don't know enough about Solr whether this is possible...Can we setup two indexes in the same core, one for product_catalog and the other for inventory? Then using a Solr query we could join the indexed content together. In Sql it would look like this select p.brand , p.model_name , p.sku , p.color_name , i.variant_id , i.variant_count from product_catalog p join inventory i on (p.sku = i.sku) On Tue, Aug 16, 2011 at 8:00 AM, Jaeger, Jay - DOT jay.jae...@dot.wi.gov wrote: On the surface, you could simply add some more fields to your schema. But as far as I can tell, you would have to have a separate Solr document for each SKU/size combination, and store the rest of the information (brand, model, color, SKU) redundantly and make the unique key a combination of the SKU and the size (presumably by having an additional field called sku_size - as far as I can tell Solr can't make a key up out of multiple fields). But, perhaps you shouldn't store that dynamic inventory information in Solr. Instead store a key that gets you to an inventory database organized by SKU and size. In such a schema, Size could be added as a multi-valued field to your existing schema, so you would know what sizes existed, so you could at least facet on that, maybe. That way Solr doesn't have to be updated every time the inventory changes. Of course, that won't help with the faceting on variant inventory. -Original Message- From: Steve Cerny [mailto:sjce...@gmail.com] Sent: Monday, August 15, 2011 6:29 PM To: solr-user@lucene.apache.org Subject: Product data schema question I'm working on an online eCommerce project and am having difficulties building the core / index schema. Here is the way we organize our product information in a normalized database. A product model has many SKUs (called colorways) A SKU has many sizes (called variants) A SKU size has associated inventory (called variant inventory) When we setup our product core we have the following field information Doc * brand * model name * SKU * color name Sample records are as follows * Haynes, Undershirt, 1234, white * Haynes, Undershirt, 1235, grey * Fruit of the Loom, Undershirt, 1236, white * Fruit of the Loom, Underwear, 1237, grey The issue I'm having is I want to add inventory to each size of each SKU for faceting. Example, SKU 1234 has sizes small, medium, large. Size small has 5 in stock, size medium 10, and size large 25. In a normalized data table I would have a separate table just for inventory and related it back to the SKU with a foreign key. How do I store size and inventory information effectively with Solr? -- Steve -- Steve -- Steve Cerny sjce...@gmail.com 715-302-0639
Re: Exact matching on names?
Hi Ron, There was a discussion about this some time back, which I implemented (with great success btw) in my own code...basically you store both the analyzed and non-analyzed versions (use string type) in the index, then send in a query like this: +name:clarke name_s:clarke^100 The name field is text so it will analyze down clarke to clark but it will match both clark and clarke and the second clause would boost the entry with clarke up to the top, which you then select with rows=1. -sujit On Tue, 2011-08-16 at 10:20 -0500, Olson, Ron wrote: Hi all- I'm missing something fundamental yet I've been unable to find the definitive answer for exact name matching. I'm indexing names using the standard text field type and my search is for the name clarke. My results include clark, which is incorrect, it needs to match clarke exactly (case insensitive). I tried textType but that doesn't work because I believe it needs to be *really* exact, whereas I'm looking for things like clark oil, bob, frank, and clark, etc. Thanks for any help, Ron DISCLAIMER: This electronic message, including any attachments, files or documents, is intended only for the addressee and may contain CONFIDENTIAL, PROPRIETARY or LEGALLY PRIVILEGED information. If you are not the intended recipient, you are hereby notified that any use, disclosure, copying or distribution of this message or any of the information included in or with it is unauthorized and strictly prohibited. If you have received this message in error, please notify the sender immediately by reply e-mail and permanently delete and destroy this message and its attachments, along with any copies thereof. This message does not create any contractual obligation on behalf of the sender or Law Bulletin Publishing Company. Thank you.
Re: Indexing from a database via SolrJ
The problem with anything automatic is that I don't see how it could know which fields in the document to map DB columns to. Unless you had fields that exactly matched column names, it would be iffy... I assume DIH actually does something like this, but don't know any way of having SolrJ automagically do this. At root these kinds of things don't generalize well, but that doesn't mean that there's not a good case for doing this. Best Erick On Tue, Aug 16, 2011 at 11:26 AM, Shawn Heisey s...@elyograg.org wrote: On 8/16/2011 7:14 AM, Erick Erickson wrote: What have you tried and what doesn't it do that you want it to do? This works, instantiating the StreamingUpdateSolrServer (server) and the JDBC connection/SQL statement are left as exercises for the readerG.: while (rs.next()) { SolrInputDocument doc = new SolrInputDocument(); String id = rs.getString(id); String title = rs.getString(title); String text = rs.getString(text); doc.addField(id, id); doc.addField(title, title); doc.addField(text, text); docs.add(doc); ++counter; ++total; if (counter 100) { // Completely arbitrary, just batch up more than one document for throughput! server.add(docs); docs.clear(); counter = 0; } } I've implemented a basic loop with the structure you've demonstrated, but it currently doesn't do anything yet with SolrInputDocument or SolrDocumentList. I figured there would be a way to avoid going through the field list one by one, but what you've written suggests that the field-by-field method is required. I can live with that. It does look like addField just takes an Object, so hopefully I can create a loop that determines the type of each field from the JDBC metadata, retrieves the correct Java type from the ResultSet, and inserts it. I imagine that everything still works if you happen to insert a field that doesn't exist in the index. This must be how the DIH does it, so I was hoping that the DIH might expose a method that takes a ResultSet and produces a SolrDocumentList. I still have to take a deeper look at the source and documentation. Thanks for the help so far, I can get a little more implemented now. Shawn
RE: Product data schema question
Not particularly. Just trying to do my part to answer some questions on the list. -Original Message- From: Steve Cerny [mailto:sjce...@gmail.com] Sent: Tuesday, August 16, 2011 11:49 AM To: solr-user@lucene.apache.org Subject: Re: Product data schema question Thanks Jay, if we come to a reasonable solution are you interested in the details? On Tue, Aug 16, 2011 at 11:44 AM, Jaeger, Jay - DOT jay.jae...@dot.wi.govwrote: No, I don't think so. A given core can only use one configuration and therefore only one schema, as far as I know, and a schema can only have one key. You could use two cores with two configurations (but that presumably wouldn't be much help). Solr is not a DBMS. It is an index. -Original Message- From: Steve Cerny [mailto:sjce...@gmail.com] Sent: Tuesday, August 16, 2011 11:37 AM To: solr-user@lucene.apache.org Subject: Re: Product data schema question Jay, this is great information. I don't know enough about Solr whether this is possible...Can we setup two indexes in the same core, one for product_catalog and the other for inventory? Then using a Solr query we could join the indexed content together. In Sql it would look like this select p.brand , p.model_name , p.sku , p.color_name , i.variant_id , i.variant_count from product_catalog p join inventory i on (p.sku = i.sku) On Tue, Aug 16, 2011 at 8:00 AM, Jaeger, Jay - DOT jay.jae...@dot.wi.gov wrote: On the surface, you could simply add some more fields to your schema. But as far as I can tell, you would have to have a separate Solr document for each SKU/size combination, and store the rest of the information (brand, model, color, SKU) redundantly and make the unique key a combination of the SKU and the size (presumably by having an additional field called sku_size - as far as I can tell Solr can't make a key up out of multiple fields). But, perhaps you shouldn't store that dynamic inventory information in Solr. Instead store a key that gets you to an inventory database organized by SKU and size. In such a schema, Size could be added as a multi-valued field to your existing schema, so you would know what sizes existed, so you could at least facet on that, maybe. That way Solr doesn't have to be updated every time the inventory changes. Of course, that won't help with the faceting on variant inventory. -Original Message- From: Steve Cerny [mailto:sjce...@gmail.com] Sent: Monday, August 15, 2011 6:29 PM To: solr-user@lucene.apache.org Subject: Product data schema question I'm working on an online eCommerce project and am having difficulties building the core / index schema. Here is the way we organize our product information in a normalized database. A product model has many SKUs (called colorways) A SKU has many sizes (called variants) A SKU size has associated inventory (called variant inventory) When we setup our product core we have the following field information Doc * brand * model name * SKU * color name Sample records are as follows * Haynes, Undershirt, 1234, white * Haynes, Undershirt, 1235, grey * Fruit of the Loom, Undershirt, 1236, white * Fruit of the Loom, Underwear, 1237, grey The issue I'm having is I want to add inventory to each size of each SKU for faceting. Example, SKU 1234 has sizes small, medium, large. Size small has 5 in stock, size medium 10, and size large 25. In a normalized data table I would have a separate table just for inventory and related it back to the SKU with a foreign key. How do I store size and inventory information effectively with Solr? -- Steve -- Steve -- Steve Cerny sjce...@gmail.com 715-302-0639
Re: Exception DirectSolrSpellChecker when using spellcheck.q
Thanx. I was using a build of the day you fixed the bug :) Keep up the good work. -- View this message in context: http://lucene.472066.n3.nabble.com/Exception-DirectSolrSpellChecker-when-using-spellcheck-q-tp3249565p3259372.html Sent from the Solr - User mailing list archive at Nabble.com.
Random + Boost?
Hey guys, This might seem odd, but is it possible to use boost with random ordering? That is, documents that get boosted are more likely to appear towards the top of the ordering (I only display page 1, say 30 documents). Does that make sense? I'm assuming that random ordering is, well, really random - so then it's not possible. But I figured I'd ask. My problem is that I want to display a random assortment of documents, but unfortunately certain types of documents far outnumber other types. So a random assortment ends up with 50% type A, 50% type B, C, D, E, F. So, I was thinking I would essentially boost types B, C, D, E, F until all types are approximately evenly represented in the random assortment. (Or alternatively, if the user has an affinity for type B documents, further boost type B documents so that they're more likely to be represented than other types). Anyone know if there's a way to do something like this in Solr? Much appreciated. Thanks. -- Steven Ou | 歐偉凡 *ravn.com* | Chief Technology Officer steve...@gmail.com | +1 909-569-9880
How to connect Solr with external Zookeeper ensemble
Hey, How could I connect my solr server with external zookeeper? Thanks, Sharath
Faceted Search Patent Lawsuit - Please Read
Just wanted to make people aware of a company called Kolera that is sending threatening letters via their law firm, manatt, phelps, phillips in regard to violation of their patent U.S. Patent No. 6,275,821 titled Method and System for Executing a Guided Parametric Search. Apparently, they believe that this patent gives them the rights to faceted search technology. Consequently, anyone who uses this technology without paying licensing fees is potentially a target of litigation. They have outstanding suits with companies such as Microsoft and Ebay. Additionally, Adobe filed a counter-suit on August 10, 2011.Unfortunately, I received a letter this afternoon and we are determining how to proceed. Just thought I would make others aware of this. I'd appreciate any insight others might have regarding the issue. -- View this message in context: http://lucene.472066.n3.nabble.com/Faceted-Search-Patent-Lawsuit-Please-Read-tp3259475p3259475.html Sent from the Solr - User mailing list archive at Nabble.com.
Search by range in multivalued fields
Hi. I have a solr core with job records and one guy can work in different companies in a specific range of dateini to dateend. doc arr name=companyinimultivaluefield companyiniIBM10012005companyini companyiniAPPLE10012005companyini /arr arr name=companyendmultivaluefield companyendIBM10012005companyend companyendAPPLE10012005companyend /arr /doc Is possible to make a range query on a multivalue field over text fields. For instance something like that. companyinimultivaluefield['IBM10012005' TO *] AND companyendmultivaluefield['IBM10012005' TO *] -- View this message in context: http://lucene.472066.n3.nabble.com/Search-by-range-in-multivalued-fields-tp3259343p3259343.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: SolrJ Spatial Search
I'm sorry, i'm trying to do the same as he is. I've read your reply many times now, but i still don't know how to do this. Would somebody help me with this? Thanks a lot -- View this message in context: http://lucene.472066.n3.nabble.com/Spatial-search-with-SolrJ-3-1-How-to-tp2961136p3259456.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Random + Boost?
This might seem odd, but is it possible to use boost with random ordering? That is, documents that get boosted are more likely to appear towards the top of the ordering (I only display page 1, say 30 documents). Does that make sense? I'm assuming that random ordering is, well, really random - so then it's not possible. But I figured I'd ask. My problem is that I want to display a random assortment of documents, but unfortunately certain types of documents far outnumber other types. So a random assortment ends up with 50% type A, 50% type B, C, D, E, F. So, I was thinking I would essentially boost types B, C, D, E, F until all types are approximately evenly represented in the random assortment. (Or alternatively, if the user has an affinity for type B documents, further boost type B documents so that they're more likely to be represented than other types). Anyone know if there's a way to do something like this in Solr? Sounds like you want to achieve diversity of results. Consider using http://wiki.apache.org/solr/FieldCollapsing Alternatively you can make use of RandomSortField with function queries. http://lucene.apache.org/solr/api/org/apache/solr/schema/RandomSortField.html
Re: Faceted Search Patent Lawsuit - Please Read
On Wed, Aug 17, 2011 at 12:03 AM, LaMaze Johnson lam...@zesco.com wrote: [...] Just thought I would make others aware of this. I'd appreciate any insight others might have regarding the issue. [...] If you will permit me a moment of levity, from the perspective of someone in India, I would say, move to a non-third-world country that does not believe in Irrational Property Rights (IPR). Apply smileys as needed. Regards, Gora
Re: Random + Boost?
To make random results i'd use something related to dates and milliseconds, not boosting. lemme think about this... 2011/8/16 Ahmet Arslan iori...@yahoo.com This might seem odd, but is it possible to use boost with random ordering? That is, documents that get boosted are more likely to appear towards the top of the ordering (I only display page 1, say 30 documents). Does that make sense? I'm assuming that random ordering is, well, really random - so then it's not possible. But I figured I'd ask. My problem is that I want to display a random assortment of documents, but unfortunately certain types of documents far outnumber other types. So a random assortment ends up with 50% type A, 50% type B, C, D, E, F. So, I was thinking I would essentially boost types B, C, D, E, F until all types are approximately evenly represented in the random assortment. (Or alternatively, if the user has an affinity for type B documents, further boost type B documents so that they're more likely to be represented than other types). Anyone know if there's a way to do something like this in Solr? Sounds like you want to achieve diversity of results. Consider using http://wiki.apache.org/solr/FieldCollapsing Alternatively you can make use of RandomSortField with function queries. http://lucene.apache.org/solr/api/org/apache/solr/schema/RandomSortField.html -- *Alexei Martchenko* | *CEO* | Superdownloads ale...@superdownloads.com.br | ale...@martchenko.com.br | (11) 5083.1018/5080.3535/5080.3533
Re: Indexing from a database via SolrJ
On 8/16/2011 11:23 AM, Erick Erickson wrote: The problem with anything automatic is that I don't see how it could know which fields in the document to map DB columns to. Unless you had fields that exactly matched column names, it would be iffy... I assume DIH actually does something like this, but don't know any way of having SolrJ automagically do this. At root these kinds of things don't generalize well, but that doesn't mean that there's not a good case for doing this. In my case, the Solr field names are in perfect sync with the database field names. My DIH config doesn't mention any fields by name, it just passes them as-is and lets the schema handle everything. I'm perfectly OK with handling everything myself in my code, but if someone had already invented the wheel, no sense in designing a new one. :) Thanks for all your help, Erick. Shawn
custom filterquery
Hello, I am writing software for an e-commerce site. Different customers can have different selections of product depending on what is priced out for them, so to get the faceting counts correct I need to filter the values based on the pricing. I have written a functionquery to get the pricing, which works fine as part of the search query, but doesn't seem to be doing anything when I try to use it in a filter query. I wrote my pricing function query based on http://www.supermind.org/blog/756/how-to-write-a-custom-solr-functionquery, and I can see the parser part getting logged from the filter query, but nothing ever calls getValues on my ValueSource. If I use my function query as part of the main query, getValues is getting called. Can anyone point me in the right direction to get this working in the filter query? Jon Wagoner
Re: Random + Boost?
Thank you for the response! I'm learning much about Solr... So I think FieldCollapsing might do the trick... So if I understand correctly, I should be able to group by type A, B, C, D, E, F, sort groups randomly, sort within groups randomly, display simple format, and get an evenly distributed set of results across A, B, C, D, E, F... But I'm not sure it's exactly what I need: - If I have, in this case, 6 document types, how do I return 25 documents? From what I understand, I need to set how many groups (let's say all 6) * how many documents per group (closest would be 4) = 24 documents. Would I simply need to return more documents and just process the first 25? - Also, this appears to me to return *exactly* evenly distributed results. How do I: - Make certain types a little more likely to appear, or - If, suppose, I have only 2 documents in type F and I query for all 6 groups, at 5 documents per group (for a total of 30 documents), will the fact that type F only has 2 documents result in a query that only returns 27 documents? If so, I would want to select the missing 3 documents evenly from the other 5 types. As for RandomSortField + function queries... I'm not sure I understand how I can use that to achieve what I need :-/ -- Steven Ou | 歐偉凡 *ravn.com* | Chief Technology Officer steve...@gmail.com | +1 909-569-9880 On Tue, Aug 16, 2011 at 12:00 PM, Ahmet Arslan iori...@yahoo.com wrote: This might seem odd, but is it possible to use boost with random ordering? That is, documents that get boosted are more likely to appear towards the top of the ordering (I only display page 1, say 30 documents). Does that make sense? I'm assuming that random ordering is, well, really random - so then it's not possible. But I figured I'd ask. My problem is that I want to display a random assortment of documents, but unfortunately certain types of documents far outnumber other types. So a random assortment ends up with 50% type A, 50% type B, C, D, E, F. So, I was thinking I would essentially boost types B, C, D, E, F until all types are approximately evenly represented in the random assortment. (Or alternatively, if the user has an affinity for type B documents, further boost type B documents so that they're more likely to be represented than other types). Anyone know if there's a way to do something like this in Solr? Sounds like you want to achieve diversity of results. Consider using http://wiki.apache.org/solr/FieldCollapsing Alternatively you can make use of RandomSortField with function queries. http://lucene.apache.org/solr/api/org/apache/solr/schema/RandomSortField.html
Re: Faceted Search Patent Lawsuit - Please Read
Gora Mohanty-3 wrote: On Wed, Aug 17, 2011 at 12:03 AM, LaMaze Johnson lt;lam...@zesco.comgt; wrote: [...] Just thought I would make others aware of this. I'd appreciate any insight others might have regarding the issue. [...] If you will permit me a moment of levity, from the perspective of someone in India, I would say, move to a non-third-world country that does not believe in Irrational Property Rights (IPR). Apply smileys as needed. Regards, Gora Permission granted. I'm interested to know how all of this will play out in the courts. Faceted search seems to be a component of every e-commerce site on the web. -- View this message in context: http://lucene.472066.n3.nabble.com/Faceted-Search-Patent-Lawsuit-Please-Read-tp3259475p3259672.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Faceted Search Patent Lawsuit - Please Read
I know you mean well and are probably wondering what to do next, but such a discussion is really beyond the scope of this mailing list. Most of us aren't lawyers (I wonder if anyone here is?) and if we were, we wouldn't likely speculate in public on something that can only be decided in the courts. -Grant On Aug 16, 2011, at 2:33 PM, LaMaze Johnson wrote: Just wanted to make people aware of a company called Kolera that is sending threatening letters via their law firm, manatt, phelps, phillips in regard to violation of their patent U.S. Patent No. 6,275,821 titled Method and System for Executing a Guided Parametric Search. Apparently, they believe that this patent gives them the rights to faceted search technology. Consequently, anyone who uses this technology without paying licensing fees is potentially a target of litigation. They have outstanding suits with companies such as Microsoft and Ebay. Additionally, Adobe filed a counter-suit on August 10, 2011.Unfortunately, I received a letter this afternoon and we are determining how to proceed. Just thought I would make others aware of this. I'd appreciate any insight others might have regarding the issue. -- View this message in context: http://lucene.472066.n3.nabble.com/Faceted-Search-Patent-Lawsuit-Please-Read-tp3259475p3259475.html Sent from the Solr - User mailing list archive at Nabble.com.
Unable to get multicore working
I've been trying (unsuccessfully) to get multicore working for about a day and a half now I'm nearly at wits end and unsure what to do anymore. **Any** help would be appreciated. I've installed Solr using the solr-jetty packages on Ubuntu 10.04. The default Solr install seems to work fine. Now, I want to add three cores: live, staging, preview to be used for the various states of the site. I've created a `solr.xml` file as follows and symlinked it in to /usr/share/solr: ?xml version=1.0 encoding=UTF-8 ? solr persistent=false cores adminPath=/admin/cores core name=preview instanceDir=/home/webteam/config/search/preview dataDir=/home/webteam/preview/data / core name=staging instanceDir=/home/webteam/config/search/staging dataDir=/home/webteam/staging/data / core name=live instanceDir=/home/webteam/config/search/live dataDir=/home/webteam/live/data / /cores /solr Now, when I try to view any cores, I get a 404 - Not found. In fact, I can't even view /solr/admin/ anymore after installing that `solr.xml` file. Also, /solr/admin/cores returns an XML file, but it looks to me like there's no cores listed. The output: response lst name=responseHeader int name=status0/int int name=QTime0/int /lst lst name=status/ /response Finally, looking through the logs produced by Jetty doesn't seem to reveal any clues about what is wrong. There doesn't seem to be any errors in there, except the 404s. Long story short. I'm stuck. Any suggestions on where to go with this? David
Re: Faceted Search Patent Lawsuit - Please Read
Grant Ingersoll-2 wrote: I know you mean well and are probably wondering what to do next, but such a discussion is really beyond the scope of this mailing list. Most of us aren't lawyers (I wonder if anyone here is?) and if we were, we wouldn't likely speculate in public on something that can only be decided in the courts. -Grant We will definitely be in touch with our legal counsel. This was more a heads up in hopes that others aren't blindsided by this. -- View this message in context: http://lucene.472066.n3.nabble.com/Faceted-Search-Patent-Lawsuit-Please-Read-tp3259475p3259733.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: SolrJ Spatial Search
solrQuery.setQuery(*:*); solrQuery.addFilterQuery({!func}geodist()); solrQuery.set(sfield, store); solrQuery.set(pt, lat + , + lon); solrQuery.set(sort, geodist() asc); //disclaimer: I haven't run this - Author: https://www.packtpub.com/solr-1-4-enterprise-search-server/book -- View this message in context: http://lucene.472066.n3.nabble.com/Spatial-search-with-SolrJ-3-1-How-to-tp2961136p3259743.html Sent from the Solr - User mailing list archive at Nabble.com.
RE: Unable to get multicore working
Perhaps your admin doesn’t work because you don't have defaultCoreName=whatever-core-you-want-by-default in your cores tag? E.g.: cores adminPath=/admin/cores defaultCoreName=collection1 Perhaps this was enough to prevent it starting any cores -- I'd expect a default to be required. Also, from experience, if you add cores, and you have security turned on, you probably need to modify web.xml to teach it about them, so it maps the URLs to the right place with the right security. Finally, check your logs to make sure that Solr isn't complaining about something else (like not having a defaultCoreName, for instance) -Original Message- From: David Sauve [mailto:dnsa...@gmail.com] Sent: Tuesday, August 16, 2011 3:02 PM To: solr-user@lucene.apache.org Subject: Unable to get multicore working I've been trying (unsuccessfully) to get multicore working for about a day and a half now I'm nearly at wits end and unsure what to do anymore. **Any** help would be appreciated. I've installed Solr using the solr-jetty packages on Ubuntu 10.04. The default Solr install seems to work fine. Now, I want to add three cores: live, staging, preview to be used for the various states of the site. I've created a `solr.xml` file as follows and symlinked it in to /usr/share/solr: ?xml version=1.0 encoding=UTF-8 ? solr persistent=false cores adminPath=/admin/cores core name=preview instanceDir=/home/webteam/config/search/preview dataDir=/home/webteam/preview/data / core name=staging instanceDir=/home/webteam/config/search/staging dataDir=/home/webteam/staging/data / core name=live instanceDir=/home/webteam/config/search/live dataDir=/home/webteam/live/data / /cores /solr Now, when I try to view any cores, I get a 404 - Not found. In fact, I can't even view /solr/admin/ anymore after installing that `solr.xml` file. Also, /solr/admin/cores returns an XML file, but it looks to me like there's no cores listed. The output: response lst name=responseHeader int name=status0/int int name=QTime0/int /lst lst name=status/ /response Finally, looking through the logs produced by Jetty doesn't seem to reveal any clues about what is wrong. There doesn't seem to be any errors in there, except the 404s. Long story short. I'm stuck. Any suggestions on where to go with this? David
Re: Unable to get multicore working
Lets try something simplier. My start.jar is on \apache-solr-3.3.0\example\ Here's my local config placed in \apache-solr-3.3.0\example\solr\ ?xml version=1.0 encoding=UTF-8 ? solr persistent=true cores adminPath=/admin/cores core name=softwares01 instanceDir=softwares01 / /cores /solr Create \apache-solr-3.3.0\example\solr\softwares01\conf\ and \apache-solr-3.3.0\example\solr\softwares01\data\ http://localhost:8983/solr/ should work and so is http://localhost:8983/solr/softwares01/admin/ 2011/8/16 David Sauve dnsa...@gmail.com I've been trying (unsuccessfully) to get multicore working for about a day and a half now I'm nearly at wits end and unsure what to do anymore. **Any** help would be appreciated. I've installed Solr using the solr-jetty packages on Ubuntu 10.04. The default Solr install seems to work fine. Now, I want to add three cores: live, staging, preview to be used for the various states of the site. I've created a `solr.xml` file as follows and symlinked it in to /usr/share/solr: ?xml version=1.0 encoding=UTF-8 ? solr persistent=false cores adminPath=/admin/cores core name=preview instanceDir=/home/webteam/config/search/preview dataDir=/home/webteam/preview/data / core name=staging instanceDir=/home/webteam/config/search/staging dataDir=/home/webteam/staging/data / core name=live instanceDir=/home/webteam/config/search/live dataDir=/home/webteam/live/data / /cores /solr Now, when I try to view any cores, I get a 404 - Not found. In fact, I can't even view /solr/admin/ anymore after installing that `solr.xml` file. Also, /solr/admin/cores returns an XML file, but it looks to me like there's no cores listed. The output: response lst name=responseHeader int name=status0/int int name=QTime0/int /lst lst name=status/ /response Finally, looking through the logs produced by Jetty doesn't seem to reveal any clues about what is wrong. There doesn't seem to be any errors in there, except the 404s. Long story short. I'm stuck. Any suggestions on where to go with this? David -- *Alexei Martchenko* | *CEO* | Superdownloads ale...@superdownloads.com.br | ale...@martchenko.com.br | (11) 5083.1018/5080.3535/5080.3533
Re: Unable to get multicore working
I've installed using aptitude so I don't have an example folder (that I can find). /solr/ does work (but lists no cores) /solr/live/admin/ does not -- 404 On Tuesday, 16 August, 2011 at 1:13 PM, Alexei Martchenko wrote: Lets try something simplier. My start.jar is on \apache-solr-3.3.0\example\ Here's my local config placed in \apache-solr-3.3.0\example\solr\ ?xml version=1.0 encoding=UTF-8 ? solr persistent=true cores adminPath=/admin/cores core name=softwares01 instanceDir=softwares01 / /cores /solr Create \apache-solr-3.3.0\example\solr\softwares01\conf\ and \apache-solr-3.3.0\example\solr\softwares01\data\ http://localhost:8983/solr/ should work and so is http://localhost:8983/solr/softwares01/admin/ 2011/8/16 David Sauve dnsa...@gmail.com (mailto:dnsa...@gmail.com) I've been trying (unsuccessfully) to get multicore working for about a day and a half now I'm nearly at wits end and unsure what to do anymore. **Any** help would be appreciated. I've installed Solr using the solr-jetty packages on Ubuntu 10.04. The default Solr install seems to work fine. Now, I want to add three cores: live, staging, preview to be used for the various states of the site. I've created a `solr.xml` file as follows and symlinked it in to /usr/share/solr: ?xml version=1.0 encoding=UTF-8 ? solr persistent=false cores adminPath=/admin/cores core name=preview instanceDir=/home/webteam/config/search/preview dataDir=/home/webteam/preview/data / core name=staging instanceDir=/home/webteam/config/search/staging dataDir=/home/webteam/staging/data / core name=live instanceDir=/home/webteam/config/search/live dataDir=/home/webteam/live/data / /cores /solr Now, when I try to view any cores, I get a 404 - Not found. In fact, I can't even view /solr/admin/ anymore after installing that `solr.xml` file. Also, /solr/admin/cores returns an XML file, but it looks to me like there's no cores listed. The output: response lst name=responseHeader int name=status0/int int name=QTime0/int /lst lst name=status/ /response Finally, looking through the logs produced by Jetty doesn't seem to reveal any clues about what is wrong. There doesn't seem to be any errors in there, except the 404s. Long story short. I'm stuck. Any suggestions on where to go with this? David -- *Alexei Martchenko* | *CEO* | Superdownloads ale...@superdownloads.com.br (mailto:ale...@superdownloads.com.br) | ale...@martchenko.com.br (mailto:ale...@martchenko.com.br) | (11) 5083.1018/5080.3535/5080.3533
Re: Indexing from a database via SolrJ
On 8/16/2011 1:12 PM, Shawn Heisey wrote: On 8/16/2011 11:23 AM, Erick Erickson wrote: The problem with anything automatic is that I don't see how it could know which fields in the document to map DB columns to. Unless you had fields that exactly matched column names, it would be iffy... I assume DIH actually does something like this, but don't know any way of having SolrJ automagically do this. At root these kinds of things don't generalize well, but that doesn't mean that there's not a good case for doing this. In my case, the Solr field names are in perfect sync with the database field names. My DIH config doesn't mention any fields by name, it just passes them as-is and lets the schema handle everything. I'm perfectly OK with handling everything myself in my code, but if someone had already invented the wheel, no sense in designing a new one. :) Thanks for all your help, Erick. Here's what I've ended up with in my method that takes a ResultSet and puts the data into Solr. I have to get a testbed set up before I can actually test this code, which will take me a while. I'm inviting comment now, knowing it might have bugs. Eclipse is happy with it, but that doesn't mean it works. :) /** * Takes an SQL ResultSet and adds the documents to solr. Does it in batches * of fetchSize. * * @param rs * @throws SQLException * @throws IOException * @throws SolrServerException */ private long addResultSet(ResultSet rs) throws SQLException, SolrServerException, IOException { long count = 0; int innerCount = 0; CollectionSolrInputDocument docs = new ArrayListSolrInputDocument(); ResultSetMetaData rsm = rs.getMetaData(); int numColumns = rsm.getColumnCount(); String[] colNames = new String[numColumns + 1]; for (int i = 1; i (numColumns + 1); i++) { colNames[i] = rsm.getColumnName(i); } while (rs.next()) { count++; innerCount++; SolrInputDocument doc = new SolrInputDocument(); for (int j = 1; j (numColumns + 1); j++) { Object f; switch (rsm.getColumnType(j)) { case Types.BIGINT: { f = rs.getLong(j); break; } case Types.INTEGER: { f = rs.getInt(j); break; } case Types.DATE: { f = rs.getDate(j); break; } case Types.FLOAT: { f = rs.getFloat(j); break; } case Types.DOUBLE: { f = rs.getDouble(j); break; } case Types.TIME: { f = rs.getDate(j); break; } case Types.BOOLEAN: { f = rs.getBoolean(j); break; } default: { f = rs.getString(j); } } doc.addField(colNames[j], f); } docs.add(doc); /** * When we reach fetchSize, index the documents and reset the inner * counter. */ if (innerCount == IdxStatic.fetchSize) { solrCore.add(docs); docs.clear(); innerCount = 0; } } /** * If the outer loop ended before the inner loop reset, index the * remaining documents. */ if (innerCount != 0) { solrCore.add(docs); } return count; }
Re: Unable to get multicore working
I tried setting `defaultCoreName=admin` and that didn't seem to change anything. I also tried adding an `env-entry` for solr/home pointing to /home/webteam/config but that didn't seem to help either. The logs don't have any errors in them, besides 404 errors. On Tuesday, 16 August, 2011 at 1:10 PM, Jaeger, Jay - DOT wrote: Perhaps your admin doesn’t work because you don't have defaultCoreName=whatever-core-you-want-by-default in your cores tag? E.g.: cores adminPath=/admin/cores defaultCoreName=collection1 Perhaps this was enough to prevent it starting any cores -- I'd expect a default to be required. Also, from experience, if you add cores, and you have security turned on, you probably need to modify web.xml (http://web.xml) to teach it about them, so it maps the URLs to the right place with the right security. Finally, check your logs to make sure that Solr isn't complaining about something else (like not having a defaultCoreName, for instance) -Original Message- From: David Sauve [mailto:dnsa...@gmail.com] Sent: Tuesday, August 16, 2011 3:02 PM To: solr-user@lucene.apache.org (mailto:solr-user@lucene.apache.org) Subject: Unable to get multicore working I've been trying (unsuccessfully) to get multicore working for about a day and a half now I'm nearly at wits end and unsure what to do anymore. **Any** help would be appreciated. I've installed Solr using the solr-jetty packages on Ubuntu 10.04. The default Solr install seems to work fine. Now, I want to add three cores: live, staging, preview to be used for the various states of the site. I've created a `solr.xml` file as follows and symlinked it in to /usr/share/solr: ?xml version=1.0 encoding=UTF-8 ? solr persistent=false cores adminPath=/admin/cores core name=preview instanceDir=/home/webteam/config/search/preview dataDir=/home/webteam/preview/data / core name=staging instanceDir=/home/webteam/config/search/staging dataDir=/home/webteam/staging/data / core name=live instanceDir=/home/webteam/config/search/live dataDir=/home/webteam/live/data / /cores /solr Now, when I try to view any cores, I get a 404 - Not found. In fact, I can't even view /solr/admin/ anymore after installing that `solr.xml` file. Also, /solr/admin/cores returns an XML file, but it looks to me like there's no cores listed. The output: response lst name=responseHeader int name=status0/int int name=QTime0/int /lst lst name=status/ /response Finally, looking through the logs produced by Jetty doesn't seem to reveal any clues about what is wrong. There doesn't seem to be any errors in there, except the 404s. Long story short. I'm stuck. Any suggestions on where to go with this? David
Re: Unable to get multicore working
AFAIK you're still seeing singlecore version where is your start.jar? search for solr.xml, see how many u've got plz. 2011/8/16 David Sauve dnsa...@gmail.com I've installed using aptitude so I don't have an example folder (that I can find). /solr/ does work (but lists no cores) /solr/live/admin/ does not -- 404 On Tuesday, 16 August, 2011 at 1:13 PM, Alexei Martchenko wrote: Lets try something simplier. My start.jar is on \apache-solr-3.3.0\example\ Here's my local config placed in \apache-solr-3.3.0\example\solr\ ?xml version=1.0 encoding=UTF-8 ? solr persistent=true cores adminPath=/admin/cores core name=softwares01 instanceDir=softwares01 / /cores /solr Create \apache-solr-3.3.0\example\solr\softwares01\conf\ and \apache-solr-3.3.0\example\solr\softwares01\data\ http://localhost:8983/solr/ should work and so is http://localhost:8983/solr/softwares01/admin/ 2011/8/16 David Sauve dnsa...@gmail.com (mailto:dnsa...@gmail.com) I've been trying (unsuccessfully) to get multicore working for about a day and a half now I'm nearly at wits end and unsure what to do anymore. **Any** help would be appreciated. I've installed Solr using the solr-jetty packages on Ubuntu 10.04. The default Solr install seems to work fine. Now, I want to add three cores: live, staging, preview to be used for the various states of the site. I've created a `solr.xml` file as follows and symlinked it in to /usr/share/solr: ?xml version=1.0 encoding=UTF-8 ? solr persistent=false cores adminPath=/admin/cores core name=preview instanceDir=/home/webteam/config/search/preview dataDir=/home/webteam/preview/data / core name=staging instanceDir=/home/webteam/config/search/staging dataDir=/home/webteam/staging/data / core name=live instanceDir=/home/webteam/config/search/live dataDir=/home/webteam/live/data / /cores /solr Now, when I try to view any cores, I get a 404 - Not found. In fact, I can't even view /solr/admin/ anymore after installing that `solr.xml` file. Also, /solr/admin/cores returns an XML file, but it looks to me like there's no cores listed. The output: response lst name=responseHeader int name=status0/int int name=QTime0/int /lst lst name=status/ /response Finally, looking through the logs produced by Jetty doesn't seem to reveal any clues about what is wrong. There doesn't seem to be any errors in there, except the 404s. Long story short. I'm stuck. Any suggestions on where to go with this? David -- *Alexei Martchenko* | *CEO* | Superdownloads ale...@superdownloads.com.br (mailto:ale...@superdownloads.com.br) | ale...@martchenko.com.br (mailto:ale...@martchenko.com.br) | (11) 5083.1018/5080.3535/5080.3533 -- *Alexei Martchenko* | *CEO* | Superdownloads ale...@superdownloads.com.br | ale...@martchenko.com.br | (11) 5083.1018/5080.3535/5080.3533
Re: Unable to get multicore working
Just the one `solr.xml`. The one I added (well, symlinked form my config folder -- I like to keep my configurations files organized so they can be managed by git) `start.jar` is in `usr/share/jetty/start.jar`. On Tuesday, 16 August, 2011 at 1:33 PM, Alexei Martchenko wrote: AFAIK you're still seeing singlecore version where is your start.jar? search for solr.xml, see how many u've got plz. 2011/8/16 David Sauve dnsa...@gmail.com (mailto:dnsa...@gmail.com) I've installed using aptitude so I don't have an example folder (that I can find). /solr/ does work (but lists no cores) /solr/live/admin/ does not -- 404 On Tuesday, 16 August, 2011 at 1:13 PM, Alexei Martchenko wrote: Lets try something simplier. My start.jar is on \apache-solr-3.3.0\example\ Here's my local config placed in \apache-solr-3.3.0\example\solr\ ?xml version=1.0 encoding=UTF-8 ? solr persistent=true cores adminPath=/admin/cores core name=softwares01 instanceDir=softwares01 / /cores /solr Create \apache-solr-3.3.0\example\solr\softwares01\conf\ and \apache-solr-3.3.0\example\solr\softwares01\data\ http://localhost:8983/solr/ should work and so is http://localhost:8983/solr/softwares01/admin/ 2011/8/16 David Sauve dnsa...@gmail.com (mailto:dnsa...@gmail.com) I've been trying (unsuccessfully) to get multicore working for about a day and a half now I'm nearly at wits end and unsure what to do anymore. **Any** help would be appreciated. I've installed Solr using the solr-jetty packages on Ubuntu 10.04. The default Solr install seems to work fine. Now, I want to add three cores: live, staging, preview to be used for the various states of the site. I've created a `solr.xml` file as follows and symlinked it in to /usr/share/solr: ?xml version=1.0 encoding=UTF-8 ? solr persistent=false cores adminPath=/admin/cores core name=preview instanceDir=/home/webteam/config/search/preview dataDir=/home/webteam/preview/data / core name=staging instanceDir=/home/webteam/config/search/staging dataDir=/home/webteam/staging/data / core name=live instanceDir=/home/webteam/config/search/live dataDir=/home/webteam/live/data / /cores /solr Now, when I try to view any cores, I get a 404 - Not found. In fact, I can't even view /solr/admin/ anymore after installing that `solr.xml` file. Also, /solr/admin/cores returns an XML file, but it looks to me like there's no cores listed. The output: response lst name=responseHeader int name=status0/int int name=QTime0/int /lst lst name=status/ /response Finally, looking through the logs produced by Jetty doesn't seem to reveal any clues about what is wrong. There doesn't seem to be any errors in there, except the 404s. Long story short. I'm stuck. Any suggestions on where to go with this? David -- *Alexei Martchenko* | *CEO* | Superdownloads ale...@superdownloads.com.br (mailto:ale...@superdownloads.com.br) | ale...@martchenko.com.br (mailto:ale...@martchenko.com.br) | (11) 5083.1018/5080.3535/5080.3533 -- *Alexei Martchenko* | *CEO* | Superdownloads ale...@superdownloads.com.br (mailto:ale...@superdownloads.com.br) | ale...@martchenko.com.br (mailto:ale...@martchenko.com.br) | (11) 5083.1018/5080.3535/5080.3533
Re: Unable to get multicore working
Is your solr.xml in usr/share/jetty/solr/solr.xml? lets try this xml instead ?xml version=1.0 encoding=UTF-8 ? solr persistent=true cores adminPath=/admin/cores core name=core01 instanceDir=core01 / core name=core02 instanceDir=core02 / core name=core03 instanceDir=core03 / /cores /solr Can you see the logs? You should see something like this 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr/'* 16/08/2011 17:30:55 org.apache.solr.servlet.SolrDispatchFilter init INFO: SolrDispatchFilter.init() 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome INFO: JNDI not configured for solr (NoInitialContextEx) 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome *INFO: solr home defaulted to 'solr/' (could not find system property or JNDI)* 16/08/2011 17:30:55 org.apache.solr.core.CoreContainer$Initializer initialize *INFO: looking for solr.xml: usr/share/jetty/solr/solr.xml* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome INFO: JNDI not configured for solr (NoInitialContextEx) 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome *INFO: solr home defaulted to 'solr/' (could not find system property or JNDI)* 16/08/2011 17:30:55 org.apache.solr.core.CoreContainer init *INFO: New CoreContainer: solrHome=solr/ instance=21357269* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr/'* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr\core01\'* 2011/8/16 David Sauve dnsa...@gmail.com Just the one `solr.xml`. The one I added (well, symlinked form my config folder -- I like to keep my configurations files organized so they can be managed by git) `start.jar` is in `usr/share/jetty/start.jar`. On Tuesday, 16 August, 2011 at 1:33 PM, Alexei Martchenko wrote: AFAIK you're still seeing singlecore version where is your start.jar? search for solr.xml, see how many u've got plz. 2011/8/16 David Sauve dnsa...@gmail.com (mailto:dnsa...@gmail.com) I've installed using aptitude so I don't have an example folder (that I can find). /solr/ does work (but lists no cores) /solr/live/admin/ does not -- 404 On Tuesday, 16 August, 2011 at 1:13 PM, Alexei Martchenko wrote: Lets try something simplier. My start.jar is on \apache-solr-3.3.0\example\ Here's my local config placed in \apache-solr-3.3.0\example\solr\ ?xml version=1.0 encoding=UTF-8 ? solr persistent=true cores adminPath=/admin/cores core name=softwares01 instanceDir=softwares01 / /cores /solr Create \apache-solr-3.3.0\example\solr\softwares01\conf\ and \apache-solr-3.3.0\example\solr\softwares01\data\ http://localhost:8983/solr/ should work and so is http://localhost:8983/solr/softwares01/admin/ 2011/8/16 David Sauve dnsa...@gmail.com (mailto:dnsa...@gmail.com) I've been trying (unsuccessfully) to get multicore working for about a day and a half now I'm nearly at wits end and unsure what to do anymore. **Any** help would be appreciated. I've installed Solr using the solr-jetty packages on Ubuntu 10.04. The default Solr install seems to work fine. Now, I want to add three cores: live, staging, preview to be used for the various states of the site. I've created a `solr.xml` file as follows and symlinked it in to /usr/share/solr: ?xml version=1.0 encoding=UTF-8 ? solr persistent=false cores adminPath=/admin/cores core name=preview instanceDir=/home/webteam/config/search/preview dataDir=/home/webteam/preview/data / core name=staging instanceDir=/home/webteam/config/search/staging dataDir=/home/webteam/staging/data / core name=live instanceDir=/home/webteam/config/search/live dataDir=/home/webteam/live/data / /cores /solr Now, when I try to view any cores, I get a 404 - Not found. In fact, I can't even view /solr/admin/ anymore after installing that `solr.xml` file. Also, /solr/admin/cores returns an XML file, but it looks to me like there's no cores listed. The output: response lst name=responseHeader int name=status0/int int name=QTime0/int /lst lst name=status/ /response Finally, looking through the logs produced by Jetty doesn't seem to reveal any clues about what is wrong. There doesn't seem to be any errors in there, except the 404s. Long story short. I'm stuck. Any suggestions on where to go with this? David -- *Alexei Martchenko* | *CEO* | Superdownloads ale...@superdownloads.com.br (mailto:ale...@superdownloads.com.br) | ale...@martchenko.com.br (mailto:ale...@martchenko.com.br) | (11)
RE: Unable to get multicore working
That won't work -- it would have to identify one of the three cores in your cores list (say, live). -Original Message- From: David Sauve [mailto:dnsa...@gmail.com] Sent: Tuesday, August 16, 2011 3:29 PM To: solr-user@lucene.apache.org Subject: Re: Unable to get multicore working I tried setting `defaultCoreName=admin` and that didn't seem to change anything. I also tried adding an `env-entry` for solr/home pointing to /home/webteam/config but that didn't seem to help either. The logs don't have any errors in them, besides 404 errors. On Tuesday, 16 August, 2011 at 1:10 PM, Jaeger, Jay - DOT wrote: Perhaps your admin doesn’t work because you don't have defaultCoreName=whatever-core-you-want-by-default in your cores tag? E.g.: cores adminPath=/admin/cores defaultCoreName=collection1 Perhaps this was enough to prevent it starting any cores -- I'd expect a default to be required. Also, from experience, if you add cores, and you have security turned on, you probably need to modify web.xml (http://web.xml) to teach it about them, so it maps the URLs to the right place with the right security. Finally, check your logs to make sure that Solr isn't complaining about something else (like not having a defaultCoreName, for instance) -Original Message- From: David Sauve [mailto:dnsa...@gmail.com] Sent: Tuesday, August 16, 2011 3:02 PM To: solr-user@lucene.apache.org (mailto:solr-user@lucene.apache.org) Subject: Unable to get multicore working I've been trying (unsuccessfully) to get multicore working for about a day and a half now I'm nearly at wits end and unsure what to do anymore. **Any** help would be appreciated. I've installed Solr using the solr-jetty packages on Ubuntu 10.04. The default Solr install seems to work fine. Now, I want to add three cores: live, staging, preview to be used for the various states of the site. I've created a `solr.xml` file as follows and symlinked it in to /usr/share/solr: ?xml version=1.0 encoding=UTF-8 ? solr persistent=false cores adminPath=/admin/cores core name=preview instanceDir=/home/webteam/config/search/preview dataDir=/home/webteam/preview/data / core name=staging instanceDir=/home/webteam/config/search/staging dataDir=/home/webteam/staging/data / core name=live instanceDir=/home/webteam/config/search/live dataDir=/home/webteam/live/data / /cores /solr Now, when I try to view any cores, I get a 404 - Not found. In fact, I can't even view /solr/admin/ anymore after installing that `solr.xml` file. Also, /solr/admin/cores returns an XML file, but it looks to me like there's no cores listed. The output: response lst name=responseHeader int name=status0/int int name=QTime0/int /lst lst name=status/ /response Finally, looking through the logs produced by Jetty doesn't seem to reveal any clues about what is wrong. There doesn't seem to be any errors in there, except the 404s. Long story short. I'm stuck. Any suggestions on where to go with this? David
Re: Unable to get multicore working
Nope. Only thing in the log: 1 [main] INFO org.mortbay.log - Logging to org.slf4j.impl.SimpleLogger(org.mortbay.log) via org.mortbay.log.Slf4jLog 173 [main] INFO org.mortbay.log - Redirecting stderr/stdout to /var/log/jetty/2011_08_16.stderrout.log On Tuesday, 16 August, 2011 at 1:45 PM, Alexei Martchenko wrote: Is your solr.xml in usr/share/jetty/solr/solr.xml? lets try this xml instead ?xml version=1.0 encoding=UTF-8 ? solr persistent=true cores adminPath=/admin/cores core name=core01 instanceDir=core01 / core name=core02 instanceDir=core02 / core name=core03 instanceDir=core03 / /cores /solr Can you see the logs? You should see something like this 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr/'* 16/08/2011 17:30:55 org.apache.solr.servlet.SolrDispatchFilter init INFO: SolrDispatchFilter.init() 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome INFO: JNDI not configured for solr (NoInitialContextEx) 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome *INFO: solr home defaulted to 'solr/' (could not find system property or JNDI)* 16/08/2011 17:30:55 org.apache.solr.core.CoreContainer$Initializer initialize *INFO: looking for solr.xml: usr/share/jetty/solr/solr.xml* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome INFO: JNDI not configured for solr (NoInitialContextEx) 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome *INFO: solr home defaulted to 'solr/' (could not find system property or JNDI)* 16/08/2011 17:30:55 org.apache.solr.core.CoreContainer init *INFO: New CoreContainer: solrHome=solr/ instance=21357269* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr/'* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr\core01\'* 2011/8/16 David Sauve dnsa...@gmail.com (mailto:dnsa...@gmail.com) Just the one `solr.xml`. The one I added (well, symlinked form my config folder -- I like to keep my configurations files organized so they can be managed by git) `start.jar` is in `usr/share/jetty/start.jar`. On Tuesday, 16 August, 2011 at 1:33 PM, Alexei Martchenko wrote: AFAIK you're still seeing singlecore version where is your start.jar? search for solr.xml, see how many u've got plz. 2011/8/16 David Sauve dnsa...@gmail.com (mailto:dnsa...@gmail.com) I've installed using aptitude so I don't have an example folder (that I can find). /solr/ does work (but lists no cores) /solr/live/admin/ does not -- 404 On Tuesday, 16 August, 2011 at 1:13 PM, Alexei Martchenko wrote: Lets try something simplier. My start.jar is on \apache-solr-3.3.0\example\ Here's my local config placed in \apache-solr-3.3.0\example\solr\ ?xml version=1.0 encoding=UTF-8 ? solr persistent=true cores adminPath=/admin/cores core name=softwares01 instanceDir=softwares01 / /cores /solr Create \apache-solr-3.3.0\example\solr\softwares01\conf\ and \apache-solr-3.3.0\example\solr\softwares01\data\ http://localhost:8983/solr/ should work and so is http://localhost:8983/solr/softwares01/admin/ 2011/8/16 David Sauve dnsa...@gmail.com (mailto:dnsa...@gmail.com) I've been trying (unsuccessfully) to get multicore working for about a day and a half now I'm nearly at wits end and unsure what to do anymore. **Any** help would be appreciated. I've installed Solr using the solr-jetty packages on Ubuntu 10.04. The default Solr install seems to work fine. Now, I want to add three cores: live, staging, preview to be used for the various states of the site. I've created a `solr.xml` file as follows and symlinked it in to /usr/share/solr: ?xml version=1.0 encoding=UTF-8 ? solr persistent=false cores adminPath=/admin/cores core name=preview instanceDir=/home/webteam/config/search/preview dataDir=/home/webteam/preview/data / core name=staging instanceDir=/home/webteam/config/search/staging dataDir=/home/webteam/staging/data / core name=live instanceDir=/home/webteam/config/search/live dataDir=/home/webteam/live/data / /cores /solr Now, when I try to view any cores, I get a 404 - Not found. In fact, I can't even view /solr/admin/ anymore after installing that `solr.xml` file. Also, /solr/admin/cores returns an XML file, but it looks to me like there's no cores listed. The output: response lst name=responseHeader int name=status0/int int name=QTime0/int /lst lst name=status/ /response
Re: Faceted Search Patent Lawsuit - Please Read
While I agree with Grant we shouldn't engage on a legal discussion, it may be worth that this thread shares a few dates of when faceted search was used in the old times... paul Le 16 août 2011 à 22:02, LaMaze Johnson a écrit : Grant Ingersoll-2 wrote: I know you mean well and are probably wondering what to do next, but such a discussion is really beyond the scope of this mailing list. Most of us aren't lawyers (I wonder if anyone here is?) and if we were, we wouldn't likely speculate in public on something that can only be decided in the courts. -Grant We will definitely be in touch with our legal counsel. This was more a heads up in hopes that others aren't blindsided by this. -- View this message in context: http://lucene.472066.n3.nabble.com/Faceted-Search-Patent-Lawsuit-Please-Read-tp3259475p3259733.html Sent from the Solr - User mailing list archive at Nabble.com.
RE: Unable to get multicore working
I tried on my own test environment -- pulling out the default core parameter out, under Solr 3.1 I got exactly your symptom: an error 404. HTTP ERROR 404 Problem accessing /solr/admin/index.jsp. Reason: missing core name in path The log showed: 2011-08-16 16:00:12.469:WARN::/solr/admin/ java.lang.IllegalStateException: STREAM at org.mortbay.jetty.Response.getWriter(Response.java:616) at org.apache.jasper.runtime.JspWriterImpl.initOut(JspWriterImpl.java:187) at org.apache.jasper.runtime.JspWriterImpl.flushBuffer(JspWriterImpl.java:180) at org.apache.jasper.runtime.PageContextImpl.release(PageContextImpl.java:237) at org.apache.jasper.runtime.JspFactoryImpl.internalReleasePageContext(JspFactoryImpl.java:173) at org.apache.jasper.runtime.JspFactoryImpl.releasePageContext(JspFactoryImpl.java:124) (etc.) Adding the defaultCoreName fixed it. I expect this is indeed your problem. -Original Message- From: David Sauve [mailto:dnsa...@gmail.com] Sent: Tuesday, August 16, 2011 3:50 PM To: solr-user@lucene.apache.org Subject: Re: Unable to get multicore working Nope. Only thing in the log: 1 [main] INFO org.mortbay.log - Logging to org.slf4j.impl.SimpleLogger(org.mortbay.log) via org.mortbay.log.Slf4jLog 173 [main] INFO org.mortbay.log - Redirecting stderr/stdout to /var/log/jetty/2011_08_16.stderrout.log On Tuesday, 16 August, 2011 at 1:45 PM, Alexei Martchenko wrote: Is your solr.xml in usr/share/jetty/solr/solr.xml? lets try this xml instead ?xml version=1.0 encoding=UTF-8 ? solr persistent=true cores adminPath=/admin/cores core name=core01 instanceDir=core01 / core name=core02 instanceDir=core02 / core name=core03 instanceDir=core03 / /cores /solr Can you see the logs? You should see something like this 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr/'* 16/08/2011 17:30:55 org.apache.solr.servlet.SolrDispatchFilter init INFO: SolrDispatchFilter.init() 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome INFO: JNDI not configured for solr (NoInitialContextEx) 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome *INFO: solr home defaulted to 'solr/' (could not find system property or JNDI)* 16/08/2011 17:30:55 org.apache.solr.core.CoreContainer$Initializer initialize *INFO: looking for solr.xml: usr/share/jetty/solr/solr.xml* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome INFO: JNDI not configured for solr (NoInitialContextEx) 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome *INFO: solr home defaulted to 'solr/' (could not find system property or JNDI)* 16/08/2011 17:30:55 org.apache.solr.core.CoreContainer init *INFO: New CoreContainer: solrHome=solr/ instance=21357269* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr/'* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr\core01\'* 2011/8/16 David Sauve dnsa...@gmail.com (mailto:dnsa...@gmail.com) Just the one `solr.xml`. The one I added (well, symlinked form my config folder -- I like to keep my configurations files organized so they can be managed by git) `start.jar` is in `usr/share/jetty/start.jar`. On Tuesday, 16 August, 2011 at 1:33 PM, Alexei Martchenko wrote: AFAIK you're still seeing singlecore version where is your start.jar? search for solr.xml, see how many u've got plz. 2011/8/16 David Sauve dnsa...@gmail.com (mailto:dnsa...@gmail.com) I've installed using aptitude so I don't have an example folder (that I can find). /solr/ does work (but lists no cores) /solr/live/admin/ does not -- 404 On Tuesday, 16 August, 2011 at 1:13 PM, Alexei Martchenko wrote: Lets try something simplier. My start.jar is on \apache-solr-3.3.0\example\ Here's my local config placed in \apache-solr-3.3.0\example\solr\ ?xml version=1.0 encoding=UTF-8 ? solr persistent=true cores adminPath=/admin/cores core name=softwares01 instanceDir=softwares01 / /cores /solr Create \apache-solr-3.3.0\example\solr\softwares01\conf\ and \apache-solr-3.3.0\example\solr\softwares01\data\ http://localhost:8983/solr/ should work and so is http://localhost:8983/solr/softwares01/admin/ 2011/8/16 David Sauve dnsa...@gmail.com (mailto:dnsa...@gmail.com) I've been trying (unsuccessfully) to get multicore working for about a day and a half now I'm nearly at wits end and unsure what to do anymore. **Any** help would be appreciated. I've installed Solr using the solr-jetty packages on Ubuntu 10.04. The default Solr install seems to work fine.
RE: Unable to get multicore working
Whoops: That was Solr 4.0 (which pre-dates 3.1). I doubt very much that the release matters, though: I expect the behavior would be the same. -Original Message- From: Jaeger, Jay - DOT [mailto:jay.jae...@dot.wi.gov] Sent: Tuesday, August 16, 2011 4:04 PM To: solr-user@lucene.apache.org Subject: RE: Unable to get multicore working I tried on my own test environment -- pulling out the default core parameter out, under Solr 3.1 I got exactly your symptom: an error 404. HTTP ERROR 404 Problem accessing /solr/admin/index.jsp. Reason: missing core name in path The log showed: 2011-08-16 16:00:12.469:WARN::/solr/admin/ java.lang.IllegalStateException: STREAM at org.mortbay.jetty.Response.getWriter(Response.java:616) at org.apache.jasper.runtime.JspWriterImpl.initOut(JspWriterImpl.java:187) at org.apache.jasper.runtime.JspWriterImpl.flushBuffer(JspWriterImpl.java:180) at org.apache.jasper.runtime.PageContextImpl.release(PageContextImpl.java:237) at org.apache.jasper.runtime.JspFactoryImpl.internalReleasePageContext(JspFactoryImpl.java:173) at org.apache.jasper.runtime.JspFactoryImpl.releasePageContext(JspFactoryImpl.java:124) (etc.) Adding the defaultCoreName fixed it. I expect this is indeed your problem. -Original Message- From: David Sauve [mailto:dnsa...@gmail.com] Sent: Tuesday, August 16, 2011 3:50 PM To: solr-user@lucene.apache.org Subject: Re: Unable to get multicore working Nope. Only thing in the log: 1 [main] INFO org.mortbay.log - Logging to org.slf4j.impl.SimpleLogger(org.mortbay.log) via org.mortbay.log.Slf4jLog 173 [main] INFO org.mortbay.log - Redirecting stderr/stdout to /var/log/jetty/2011_08_16.stderrout.log On Tuesday, 16 August, 2011 at 1:45 PM, Alexei Martchenko wrote: Is your solr.xml in usr/share/jetty/solr/solr.xml? lets try this xml instead ?xml version=1.0 encoding=UTF-8 ? solr persistent=true cores adminPath=/admin/cores core name=core01 instanceDir=core01 / core name=core02 instanceDir=core02 / core name=core03 instanceDir=core03 / /cores /solr Can you see the logs? You should see something like this 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr/'* 16/08/2011 17:30:55 org.apache.solr.servlet.SolrDispatchFilter init INFO: SolrDispatchFilter.init() 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome INFO: JNDI not configured for solr (NoInitialContextEx) 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome *INFO: solr home defaulted to 'solr/' (could not find system property or JNDI)* 16/08/2011 17:30:55 org.apache.solr.core.CoreContainer$Initializer initialize *INFO: looking for solr.xml: usr/share/jetty/solr/solr.xml* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome INFO: JNDI not configured for solr (NoInitialContextEx) 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome *INFO: solr home defaulted to 'solr/' (could not find system property or JNDI)* 16/08/2011 17:30:55 org.apache.solr.core.CoreContainer init *INFO: New CoreContainer: solrHome=solr/ instance=21357269* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr/'* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr\core01\'* 2011/8/16 David Sauve dnsa...@gmail.com (mailto:dnsa...@gmail.com) Just the one `solr.xml`. The one I added (well, symlinked form my config folder -- I like to keep my configurations files organized so they can be managed by git) `start.jar` is in `usr/share/jetty/start.jar`. On Tuesday, 16 August, 2011 at 1:33 PM, Alexei Martchenko wrote: AFAIK you're still seeing singlecore version where is your start.jar? search for solr.xml, see how many u've got plz. 2011/8/16 David Sauve dnsa...@gmail.com (mailto:dnsa...@gmail.com) I've installed using aptitude so I don't have an example folder (that I can find). /solr/ does work (but lists no cores) /solr/live/admin/ does not -- 404 On Tuesday, 16 August, 2011 at 1:13 PM, Alexei Martchenko wrote: Lets try something simplier. My start.jar is on \apache-solr-3.3.0\example\ Here's my local config placed in \apache-solr-3.3.0\example\solr\ ?xml version=1.0 encoding=UTF-8 ? solr persistent=true cores adminPath=/admin/cores core name=softwares01 instanceDir=softwares01 / /cores /solr Create \apache-solr-3.3.0\example\solr\softwares01\conf\ and \apache-solr-3.3.0\example\solr\softwares01\data\ http://localhost:8983/solr/ should work and so is http://localhost:8983/solr/softwares01/admin/ 2011/8/16 David Sauve dnsa...@gmail.com (mailto:dnsa...@gmail.com) I've
Re: How to deal with java.net.SocketTimeoutException: Read timed out on commit?
We had this type of error too. Now we are using the StreamingUpdateSolrServer with a quite big queue and 2-4 threads depending on data type: http://lucene.apache.org/solr/api/org/apache/solr/client/solrj/impl/StreamingUpdateSolrServer.html And we do not do any intermediate commit. We send only one commit at the end of the process and now everything is fine. Ludovic. - Jouve France. -- View this message in context: http://lucene.472066.n3.nabble.com/How-to-deal-with-java-net-SocketTimeoutException-Read-timed-out-on-commit-tp3258430p3259890.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Unable to get multicore working
I updated my `solr.xml` as follow: ?xml version=1.0 encoding=UTF-8 ? solr persistent=false cores adminPath=/admin/cores defaultCoreName=live core name=preview instanceDir=/home/webteam/config/search/preview dataDir=/home/webteam/preview/data / core name=staging instanceDir=/home/webteam/config/search/staging dataDir=/home/webteam/staging/data / core name=live instanceDir=/home/webteam/config/search/live dataDir=/home/webteam/live/data / /cores /solr and I'm still seeing the same 404 when I true to view /solr/admin/ or /solr/live/admin/ That said, the logs are showing a different error now. Excellent! The site schemas are loading! Looks like the site schemas have an issue: SEVERE: org.apache.solr.common.SolrException: Unknown fieldtype 'long' specified on field area_id Errr. Why would `long` be an invalid type? On Tuesday, 16 August, 2011 at 2:06 PM, Jaeger, Jay - DOT wrote: Whoops: That was Solr 4.0 (which pre-dates 3.1). I doubt very much that the release matters, though: I expect the behavior would be the same. -Original Message- From: Jaeger, Jay - DOT [mailto:jay.jae...@dot.wi.gov] Sent: Tuesday, August 16, 2011 4:04 PM To: solr-user@lucene.apache.org (mailto:solr-user@lucene.apache.org) Subject: RE: Unable to get multicore working I tried on my own test environment -- pulling out the default core parameter out, under Solr 3.1 I got exactly your symptom: an error 404. HTTP ERROR 404 Problem accessing /solr/admin/index.jsp. Reason: missing core name in path The log showed: 2011-08-16 16:00:12.469:WARN::/solr/admin/ java.lang.IllegalStateException: STREAM at org.mortbay.jetty.Response.getWriter(Response.java:616) at org.apache.jasper.runtime.JspWriterImpl.initOut(JspWriterImpl.java:187) at org.apache.jasper.runtime.JspWriterImpl.flushBuffer(JspWriterImpl.java:180) at org.apache.jasper.runtime.PageContextImpl.release(PageContextImpl.java:237) at org.apache.jasper.runtime.JspFactoryImpl.internalReleasePageContext(JspFactoryImpl.java:173) at org.apache.jasper.runtime.JspFactoryImpl.releasePageContext(JspFactoryImpl.java:124) (etc.) Adding the defaultCoreName fixed it. I expect this is indeed your problem. -Original Message- From: David Sauve [mailto:dnsa...@gmail.com] Sent: Tuesday, August 16, 2011 3:50 PM To: solr-user@lucene.apache.org (mailto:solr-user@lucene.apache.org) Subject: Re: Unable to get multicore working Nope. Only thing in the log: 1 [main] INFO org.mortbay.log - Logging to org.slf4j.impl.SimpleLogger(org.mortbay.log) via org.mortbay.log.Slf4jLog 173 [main] INFO org.mortbay.log - Redirecting stderr/stdout to /var/log/jetty/2011_08_16.stderrout.log On Tuesday, 16 August, 2011 at 1:45 PM, Alexei Martchenko wrote: Is your solr.xml in usr/share/jetty/solr/solr.xml? lets try this xml instead ?xml version=1.0 encoding=UTF-8 ? solr persistent=true cores adminPath=/admin/cores core name=core01 instanceDir=core01 / core name=core02 instanceDir=core02 / core name=core03 instanceDir=core03 / /cores /solr Can you see the logs? You should see something like this 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr/'* 16/08/2011 17:30:55 org.apache.solr.servlet.SolrDispatchFilter init INFO: SolrDispatchFilter.init() 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome INFO: JNDI not configured for solr (NoInitialContextEx) 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome *INFO: solr home defaulted to 'solr/' (could not find system property or JNDI)* 16/08/2011 17:30:55 org.apache.solr.core.CoreContainer$Initializer initialize *INFO: looking for solr.xml: usr/share/jetty/solr/solr.xml* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome INFO: JNDI not configured for solr (NoInitialContextEx) 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome *INFO: solr home defaulted to 'solr/' (could not find system property or JNDI)* 16/08/2011 17:30:55 org.apache.solr.core.CoreContainer init *INFO: New CoreContainer: solrHome=solr/ instance=21357269* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr/'* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr\core01\'* 2011/8/16 David Sauve dnsa...@gmail.com (mailto:dnsa...@gmail.com) Just the one `solr.xml`. The one I added (well, symlinked form my config folder -- I like to keep my configurations files organized so they can be managed by git) `start.jar` is in `usr/share/jetty/start.jar`. On Tuesday, 16 August, 2011 at 1:33 PM, Alexei Martchenko wrote: AFAIK you're still seeing singlecore version where is your start.jar? search for solr.xml, see how many u've got plz.
Re: Unable to get multicore working
When you go to /solr what do you see? On Tue, Aug 16, 2011 at 5:23 PM, David Sauve dnsa...@gmail.com wrote: I updated my `solr.xml` as follow: ?xml version=1.0 encoding=UTF-8 ? solr persistent=false cores adminPath=/admin/cores defaultCoreName=live core name=preview instanceDir=/home/webteam/config/search/preview dataDir=/home/webteam/preview/data / core name=staging instanceDir=/home/webteam/config/search/staging dataDir=/home/webteam/staging/data / core name=live instanceDir=/home/webteam/config/search/live dataDir=/home/webteam/live/data / /cores /solr and I'm still seeing the same 404 when I true to view /solr/admin/ or /solr/live/admin/ That said, the logs are showing a different error now. Excellent! The site schemas are loading! Looks like the site schemas have an issue: SEVERE: org.apache.solr.common.SolrException: Unknown fieldtype 'long' specified on field area_id Errr. Why would `long` be an invalid type? On Tuesday, 16 August, 2011 at 2:06 PM, Jaeger, Jay - DOT wrote: Whoops: That was Solr 4.0 (which pre-dates 3.1). I doubt very much that the release matters, though: I expect the behavior would be the same. -Original Message- From: Jaeger, Jay - DOT [mailto:jay.jae...@dot.wi.gov] Sent: Tuesday, August 16, 2011 4:04 PM To: solr-user@lucene.apache.org (mailto:solr-user@lucene.apache.org) Subject: RE: Unable to get multicore working I tried on my own test environment -- pulling out the default core parameter out, under Solr 3.1 I got exactly your symptom: an error 404. HTTP ERROR 404 Problem accessing /solr/admin/index.jsp. Reason: missing core name in path The log showed: 2011-08-16 16:00:12.469:WARN::/solr/admin/ java.lang.IllegalStateException: STREAM at org.mortbay.jetty.Response.getWriter(Response.java:616) at org.apache.jasper.runtime.JspWriterImpl.initOut(JspWriterImpl.java:187) at org.apache.jasper.runtime.JspWriterImpl.flushBuffer(JspWriterImpl.java:180) at org.apache.jasper.runtime.PageContextImpl.release(PageContextImpl.java:237) at org.apache.jasper.runtime.JspFactoryImpl.internalReleasePageContext(JspFactoryImpl.java:173) at org.apache.jasper.runtime.JspFactoryImpl.releasePageContext(JspFactoryImpl.java:124) (etc.) Adding the defaultCoreName fixed it. I expect this is indeed your problem. -Original Message- From: David Sauve [mailto:dnsa...@gmail.com] Sent: Tuesday, August 16, 2011 3:50 PM To: solr-user@lucene.apache.org (mailto:solr-user@lucene.apache.org) Subject: Re: Unable to get multicore working Nope. Only thing in the log: 1 [main] INFO org.mortbay.log - Logging to org.slf4j.impl.SimpleLogger(org.mortbay.log) via org.mortbay.log.Slf4jLog 173 [main] INFO org.mortbay.log - Redirecting stderr/stdout to /var/log/jetty/2011_08_16.stderrout.log On Tuesday, 16 August, 2011 at 1:45 PM, Alexei Martchenko wrote: Is your solr.xml in usr/share/jetty/solr/solr.xml? lets try this xml instead ?xml version=1.0 encoding=UTF-8 ? solr persistent=true cores adminPath=/admin/cores core name=core01 instanceDir=core01 / core name=core02 instanceDir=core02 / core name=core03 instanceDir=core03 / /cores /solr Can you see the logs? You should see something like this 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr/'* 16/08/2011 17:30:55 org.apache.solr.servlet.SolrDispatchFilter init INFO: SolrDispatchFilter.init() 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome INFO: JNDI not configured for solr (NoInitialContextEx) 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome *INFO: solr home defaulted to 'solr/' (could not find system property or JNDI)* 16/08/2011 17:30:55 org.apache.solr.core.CoreContainer$Initializer initialize *INFO: looking for solr.xml: usr/share/jetty/solr/solr.xml* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome INFO: JNDI not configured for solr (NoInitialContextEx) 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome *INFO: solr home defaulted to 'solr/' (could not find system property or JNDI)* 16/08/2011 17:30:55 org.apache.solr.core.CoreContainer init *INFO: New CoreContainer: solrHome=solr/ instance=21357269* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr/'* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr\core01\'* 2011/8/16 David Sauve dnsa...@gmail.com (mailto:dnsa...@gmail.com) Just the one `solr.xml`. The one I added (well, symlinked form my config folder -- I like to keep my configurations files organized so they can be managed by git) `start.jar` is in `usr/share/jetty/start.jar`. On Tuesday, 16 August, 2011
Re: Unable to get multicore working
Welcome to Solr with a link to Admin. The link returns a 404. On Tuesday, 16 August, 2011 at 2:30 PM, Donald Organ wrote: When you go to /solr what do you see? On Tue, Aug 16, 2011 at 5:23 PM, David Sauve dnsa...@gmail.com (mailto:dnsa...@gmail.com) wrote: I updated my `solr.xml` as follow: ?xml version=1.0 encoding=UTF-8 ? solr persistent=false cores adminPath=/admin/cores defaultCoreName=live core name=preview instanceDir=/home/webteam/config/search/preview dataDir=/home/webteam/preview/data / core name=staging instanceDir=/home/webteam/config/search/staging dataDir=/home/webteam/staging/data / core name=live instanceDir=/home/webteam/config/search/live dataDir=/home/webteam/live/data / /cores /solr and I'm still seeing the same 404 when I true to view /solr/admin/ or /solr/live/admin/ That said, the logs are showing a different error now. Excellent! The site schemas are loading! Looks like the site schemas have an issue: SEVERE: org.apache.solr.common.SolrException: Unknown fieldtype 'long' specified on field area_id Errr. Why would `long` be an invalid type? On Tuesday, 16 August, 2011 at 2:06 PM, Jaeger, Jay - DOT wrote: Whoops: That was Solr 4.0 (which pre-dates 3.1). I doubt very much that the release matters, though: I expect the behavior would be the same. -Original Message- From: Jaeger, Jay - DOT [mailto:jay.jae...@dot.wi.gov] Sent: Tuesday, August 16, 2011 4:04 PM To: solr-user@lucene.apache.org (mailto:solr-user@lucene.apache.org) Subject: RE: Unable to get multicore working I tried on my own test environment -- pulling out the default core parameter out, under Solr 3.1 I got exactly your symptom: an error 404. HTTP ERROR 404 Problem accessing /solr/admin/index.jsp. Reason: missing core name in path The log showed: 2011-08-16 16:00:12.469:WARN::/solr/admin/ java.lang.IllegalStateException: STREAM at org.mortbay.jetty.Response.getWriter(Response.java:616) at org.apache.jasper.runtime.JspWriterImpl.initOut(JspWriterImpl.java:187) at org.apache.jasper.runtime.JspWriterImpl.flushBuffer(JspWriterImpl.java:180) at org.apache.jasper.runtime.PageContextImpl.release(PageContextImpl.java:237) at org.apache.jasper.runtime.JspFactoryImpl.internalReleasePageContext(JspFactoryImpl.java:173) at org.apache.jasper.runtime.JspFactoryImpl.releasePageContext(JspFactoryImpl.java:124) (etc.) Adding the defaultCoreName fixed it. I expect this is indeed your problem. -Original Message- From: David Sauve [mailto:dnsa...@gmail.com] Sent: Tuesday, August 16, 2011 3:50 PM To: solr-user@lucene.apache.org (mailto:solr-user@lucene.apache.org) Subject: Re: Unable to get multicore working Nope. Only thing in the log: 1 [main] INFO org.mortbay.log - Logging to org.slf4j.impl.SimpleLogger(org.mortbay.log) via org.mortbay.log.Slf4jLog 173 [main] INFO org.mortbay.log - Redirecting stderr/stdout to /var/log/jetty/2011_08_16.stderrout.log On Tuesday, 16 August, 2011 at 1:45 PM, Alexei Martchenko wrote: Is your solr.xml in usr/share/jetty/solr/solr.xml? lets try this xml instead ?xml version=1.0 encoding=UTF-8 ? solr persistent=true cores adminPath=/admin/cores core name=core01 instanceDir=core01 / core name=core02 instanceDir=core02 / core name=core03 instanceDir=core03 / /cores /solr Can you see the logs? You should see something like this 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr/'* 16/08/2011 17:30:55 org.apache.solr.servlet.SolrDispatchFilter init INFO: SolrDispatchFilter.init() 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome INFO: JNDI not configured for solr (NoInitialContextEx) 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome *INFO: solr home defaulted to 'solr/' (could not find system property or JNDI)* 16/08/2011 17:30:55 org.apache.solr.core.CoreContainer$Initializer initialize *INFO: looking for solr.xml: usr/share/jetty/solr/solr.xml* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome INFO: JNDI not configured for solr (NoInitialContextEx) 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome *INFO: solr home defaulted to 'solr/' (could not find system property or JNDI)* 16/08/2011 17:30:55 org.apache.solr.core.CoreContainer init *INFO: New CoreContainer: solrHome=solr/ instance=21357269* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr/'* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr\core01\'* 2011/8/16 David
RE: Unable to get multicore working
That said, the logs are showing a different error now. Excellent! The site schemas are loading! Great! SEVERE: org.apache.solr.common.SolrException: Unknown fieldtype 'long' specified on field area_id Go have a look at your conf/schema.xml. Is the following line present?? Does your field definition for area_id follow it? fieldType name=long class=solr.TrieLongField precisionStep=0 omitNorms=true positionIncrementGap=0/ Look at the file with an XML editor. Perhaps an edit to some earlier portion of the schema is messing up this part of the schema? -Original Message- From: David Sauve [mailto:dnsa...@gmail.com] Sent: Tuesday, August 16, 2011 4:24 PM To: solr-user@lucene.apache.org Subject: Re: Unable to get multicore working I updated my `solr.xml` as follow: ?xml version=1.0 encoding=UTF-8 ? solr persistent=false cores adminPath=/admin/cores defaultCoreName=live core name=preview instanceDir=/home/webteam/config/search/preview dataDir=/home/webteam/preview/data / core name=staging instanceDir=/home/webteam/config/search/staging dataDir=/home/webteam/staging/data / core name=live instanceDir=/home/webteam/config/search/live dataDir=/home/webteam/live/data / /cores /solr and I'm still seeing the same 404 when I true to view /solr/admin/ or /solr/live/admin/ That said, the logs are showing a different error now. Excellent! The site schemas are loading! Looks like the site schemas have an issue: SEVERE: org.apache.solr.common.SolrException: Unknown fieldtype 'long' specified on field area_id Errr. Why would `long` be an invalid type? On Tuesday, 16 August, 2011 at 2:06 PM, Jaeger, Jay - DOT wrote: Whoops: That was Solr 4.0 (which pre-dates 3.1). I doubt very much that the release matters, though: I expect the behavior would be the same. -Original Message- From: Jaeger, Jay - DOT [mailto:jay.jae...@dot.wi.gov] Sent: Tuesday, August 16, 2011 4:04 PM To: solr-user@lucene.apache.org (mailto:solr-user@lucene.apache.org) Subject: RE: Unable to get multicore working I tried on my own test environment -- pulling out the default core parameter out, under Solr 3.1 I got exactly your symptom: an error 404. HTTP ERROR 404 Problem accessing /solr/admin/index.jsp. Reason: missing core name in path The log showed: 2011-08-16 16:00:12.469:WARN::/solr/admin/ java.lang.IllegalStateException: STREAM at org.mortbay.jetty.Response.getWriter(Response.java:616) at org.apache.jasper.runtime.JspWriterImpl.initOut(JspWriterImpl.java:187) at org.apache.jasper.runtime.JspWriterImpl.flushBuffer(JspWriterImpl.java:180) at org.apache.jasper.runtime.PageContextImpl.release(PageContextImpl.java:237) at org.apache.jasper.runtime.JspFactoryImpl.internalReleasePageContext(JspFactoryImpl.java:173) at org.apache.jasper.runtime.JspFactoryImpl.releasePageContext(JspFactoryImpl.java:124) (etc.) Adding the defaultCoreName fixed it. I expect this is indeed your problem. -Original Message- From: David Sauve [mailto:dnsa...@gmail.com] Sent: Tuesday, August 16, 2011 3:50 PM To: solr-user@lucene.apache.org (mailto:solr-user@lucene.apache.org) Subject: Re: Unable to get multicore working Nope. Only thing in the log: 1 [main] INFO org.mortbay.log - Logging to org.slf4j.impl.SimpleLogger(org.mortbay.log) via org.mortbay.log.Slf4jLog 173 [main] INFO org.mortbay.log - Redirecting stderr/stdout to /var/log/jetty/2011_08_16.stderrout.log On Tuesday, 16 August, 2011 at 1:45 PM, Alexei Martchenko wrote: Is your solr.xml in usr/share/jetty/solr/solr.xml? lets try this xml instead ?xml version=1.0 encoding=UTF-8 ? solr persistent=true cores adminPath=/admin/cores core name=core01 instanceDir=core01 / core name=core02 instanceDir=core02 / core name=core03 instanceDir=core03 / /cores /solr Can you see the logs? You should see something like this 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr/'* 16/08/2011 17:30:55 org.apache.solr.servlet.SolrDispatchFilter init INFO: SolrDispatchFilter.init() 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome INFO: JNDI not configured for solr (NoInitialContextEx) 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome *INFO: solr home defaulted to 'solr/' (could not find system property or JNDI)* 16/08/2011 17:30:55 org.apache.solr.core.CoreContainer$Initializer initialize *INFO: looking for solr.xml: usr/share/jetty/solr/solr.xml* 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome INFO: JNDI not configured for solr (NoInitialContextEx) 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome *INFO: solr home defaulted to 'solr/' (could not find system property or JNDI)* 16/08/2011 17:30:55 org.apache.solr.core.CoreContainer init *INFO: New CoreContainer:
Re: Unable to get multicore working
Ok. Fixed that too, now. The schema didn't define long. Looks like everything is a-okay, now. Thanks for the help. You guys saved me from the insane asylum. On Tuesday, 16 August, 2011 at 2:32 PM, Jaeger, Jay - DOT wrote: That said, the logs are showing a different error now. Excellent! The site schemas are loading! Great! SEVERE: org.apache.solr.common.SolrException: Unknown fieldtype 'long' specified on field area_id Go have a look at your conf/schema.xml. Is the following line present?? Does your field definition for area_id follow it? fieldType name=long class=solr.TrieLongField precisionStep=0 omitNorms=true positionIncrementGap=0/ Look at the file with an XML editor. Perhaps an edit to some earlier portion of the schema is messing up this part of the schema? -Original Message- From: David Sauve [mailto:dnsa...@gmail.com] Sent: Tuesday, August 16, 2011 4:24 PM To: solr-user@lucene.apache.org (mailto:solr-user@lucene.apache.org) Subject: Re: Unable to get multicore working I updated my `solr.xml` as follow: ?xml version=1.0 encoding=UTF-8 ? solr persistent=false cores adminPath=/admin/cores defaultCoreName=live core name=preview instanceDir=/home/webteam/config/search/preview dataDir=/home/webteam/preview/data / core name=staging instanceDir=/home/webteam/config/search/staging dataDir=/home/webteam/staging/data / core name=live instanceDir=/home/webteam/config/search/live dataDir=/home/webteam/live/data / /cores /solr and I'm still seeing the same 404 when I true to view /solr/admin/ or /solr/live/admin/ That said, the logs are showing a different error now. Excellent! The site schemas are loading! Looks like the site schemas have an issue: SEVERE: org.apache.solr.common.SolrException: Unknown fieldtype 'long' specified on field area_id Errr. Why would `long` be an invalid type? On Tuesday, 16 August, 2011 at 2:06 PM, Jaeger, Jay - DOT wrote: Whoops: That was Solr 4.0 (which pre-dates 3.1). I doubt very much that the release matters, though: I expect the behavior would be the same. -Original Message- From: Jaeger, Jay - DOT [mailto:jay.jae...@dot.wi.gov] Sent: Tuesday, August 16, 2011 4:04 PM To: solr-user@lucene.apache.org (mailto:solr-user@lucene.apache.org) (mailto:solr-user@lucene.apache.org) Subject: RE: Unable to get multicore working I tried on my own test environment -- pulling out the default core parameter out, under Solr 3.1 I got exactly your symptom: an error 404. HTTP ERROR 404 Problem accessing /solr/admin/index.jsp. Reason: missing core name in path The log showed: 2011-08-16 16:00:12.469:WARN::/solr/admin/ java.lang.IllegalStateException: STREAM at org.mortbay.jetty.Response.getWriter(Response.java:616) at org.apache.jasper.runtime.JspWriterImpl.initOut(JspWriterImpl.java:187) at org.apache.jasper.runtime.JspWriterImpl.flushBuffer(JspWriterImpl.java:180) at org.apache.jasper.runtime.PageContextImpl.release(PageContextImpl.java:237) at org.apache.jasper.runtime.JspFactoryImpl.internalReleasePageContext(JspFactoryImpl.java:173) at org.apache.jasper.runtime.JspFactoryImpl.releasePageContext(JspFactoryImpl.java:124) (etc.) Adding the defaultCoreName fixed it. I expect this is indeed your problem. -Original Message- From: David Sauve [mailto:dnsa...@gmail.com] Sent: Tuesday, August 16, 2011 3:50 PM To: solr-user@lucene.apache.org (mailto:solr-user@lucene.apache.org) (mailto:solr-user@lucene.apache.org) Subject: Re: Unable to get multicore working Nope. Only thing in the log: 1 [main] INFO org.mortbay.log - Logging to org.slf4j.impl.SimpleLogger(org.mortbay.log) via org.mortbay.log.Slf4jLog 173 [main] INFO org.mortbay.log - Redirecting stderr/stdout to /var/log/jetty/2011_08_16.stderrout.log On Tuesday, 16 August, 2011 at 1:45 PM, Alexei Martchenko wrote: Is your solr.xml in usr/share/jetty/solr/solr.xml? lets try this xml instead ?xml version=1.0 encoding=UTF-8 ? solr persistent=true cores adminPath=/admin/cores core name=core01 instanceDir=core01 / core name=core02 instanceDir=core02 / core name=core03 instanceDir=core03 / /cores /solr Can you see the logs? You should see something like this 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader init *INFO: Solr home set to 'solr/'* 16/08/2011 17:30:55 org.apache.solr.servlet.SolrDispatchFilter init INFO: SolrDispatchFilter.init() 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome INFO: JNDI not configured for solr (NoInitialContextEx) 16/08/2011 17:30:55 org.apache.solr.core.SolrResourceLoader locateSolrHome *INFO: solr home defaulted to 'solr/' (could not find system property or JNDI)* 16/08/2011 17:30:55
Re: Migration from Autonomy IDOL to SOLR
Hello Karsten. From the doc you provided, it seems the two are totally different products. I thought a bit about it and it seems that the best aproach would be to: 1-refactor our app and add an abstraction layer that will call the IDOL ACI API. Make sure we have good tests in place. 2-implement a solr-dialect for the abstraction layer that we can switch on or off in config. Thanks. Arcadius. On Tue, Aug 16, 2011 at 4:09 PM, karsten-s...@gmx.de wrote: Hi Arcadius, currently we have a migration project from verity k2 search server to solr. I do not know IDOL, but autonomy bought verity before IDOL was released, so possible it is comparable? verity k2 works directly on xml-Files, in result the query syntax is a little bit like xpath e.g. with text1 IN zone2 IN zone1 instead of contains(//zone1/zone2,'text1'). About verity query syntax: http://gregconely.getmyip.com/dl/OTG%20Software/5.30.087%20Suite%20%28SP3%29/Disc%204%20-%20Verity/Verity%20K2%20Server%205.5/doc/docs/pdf/VerityQueryLanguage.pdf Does IDOL work the same way? Best regards Karsten P.S. in Context: http://lucene.472066.n3.nabble.com/Migration-from-Autonomy-IDOL-to-SOLR-td3255377.html Original-Nachricht Datum: Mon, 15 Aug 2011 11:11:36 +0100 Von: Arcadius Ahouansou arcad...@menelic.com An: solr-user@lucene.apache.org Betreff: Migration from Autonomy IDOL to SOLR Hello. We have a couple of application running on half a dozen Autonomy IDOL servers. Currently, all feature we need are supported by Solr. We have done some internal testing and realized that SOLR would do a better job. So, we are investigation all possibilities for a smooth migration from IDOL to SOLR. I am looking for advice from people who went through something similar. Ideally, we would like to keep most of our legacy code unchanged and have a kind of query-translation-layer plugged into our app if possible. -Is there lib available? -Any thought? Thanks. Arcadius.
Problems generating war distribution using ant
So the way I generate war files now is by running an 'ant dist' in the solr folder. It generates the war fine and I get a build success, and then I deploy it to tomcat and once again the logs show it was successful (from the looks of it). However, when I go to 'myip:8080/solr/admin' I get an HTTP status 404. However, it works when I take a war from the nightly build, expand it, drop some new class files in there that I need, and close it up again. The solr I have checked out seems fine though and I can't find any differences between the war I'm generating and the one that has been generated. -- View this message in context: http://lucene.472066.n3.nabble.com/Problems-generating-war-distribution-using-ant-tp3260070p3260070.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Problems generating war distribution using ant
FWIW, we have some custom classes on top of solr as well. The way we do it is using the following ant target: target name=war depends=jar description=Rebuild Solr WAR with custom code mkdir dir=${maven.webapps.output}/ !-- we unwar a copy of the 3.2.0 war file in source repo -- unwar src=${prod.common.lib.external.solr}/apache-solr-3.2.0.war dest=${maven.webapps.output}/ !-- add in some extra jar files our custom stuff needs -- copy todir=${maven.webapps.output}/WEB-INF/lib fileset refid=.../ fileset refid=.../ ... /copy !-- the jar target builds just our custom classes into a hl-solr.jar, which is copied over to the WEB-INF/lib of the exploded solr war -- copy file=${maven.build.directory}/hl-solr.jar todir=${maven.webapps.output}/WEB-INF/lib/ /war Seems to work fine...basically automates what you have described in your second paragraph, but allows us to keep our own code separately from solr code under source control. -sujit On Tue, 2011-08-16 at 16:09 -0700, arian487 wrote: So the way I generate war files now is by running an 'ant dist' in the solr folder. It generates the war fine and I get a build success, and then I deploy it to tomcat and once again the logs show it was successful (from the looks of it). However, when I go to 'myip:8080/solr/admin' I get an HTTP status 404. However, it works when I take a war from the nightly build, expand it, drop some new class files in there that I need, and close it up again. The solr I have checked out seems fine though and I can't find any differences between the war I'm generating and the one that has been generated. -- View this message in context: http://lucene.472066.n3.nabble.com/Problems-generating-war-distribution-using-ant-tp3260070p3260070.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Problems generating war distribution using ant
Interesting. I can use this as an option and create a custom 'war' target if need be but I'd like to avoid this. I'd rather do a full build from the source code I have checked out from the SVN. Any reason why 'ant dist' doesn't produce a good war file? -- View this message in context: http://lucene.472066.n3.nabble.com/Problems-generating-war-distribution-using-ant-tp3260070p3260122.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Problems generating war distribution using ant
Interesting. I can use this as an option and create a custom 'war' target if need be but I'd like to avoid this. I'd rather do a full build from the source code I have checked out from the SVN. Any reason why 'ant dist' doesn't produce a good war file? -- View this message in context: http://lucene.472066.n3.nabble.com/Problems-generating-war-distribution-using-ant-tp3260070p3260126.html Sent from the Solr - User mailing list archive at Nabble.com.
Solr UIMA integration problem
Hello, I am using Solr 3.3. I have been following instructions at https://svn.apache.org/repos/asf/lucene/dev/tags/lucene_solr_3_3/solr/contrib/uima/README.txt My setup looks like the following. solr lib directory contains the following jars apache-solr-uima-3.3.0.jar commons-digester-2.0.jar uima-an-alchemy-2.3.1-SNAPSHOT-r1062868.jar uima-an-calais-2.3.1-SNAPSHOT-r1062868.jar uima-an-tagger-2.3.1-SNAPSHOT-r1062868.jar uima-an-wst-2.3.1-SNAPSHOT-r1076132.jar uimaj-core-2.3.1.jar solr_config.xml has the following changes. updateRequestProcessorChain name=uima processor class=org.apache.solr.uima.processor.UIMAUpdateRequestProcessorFactory lst name=uimaConfig lst name=runtimeParameters str name=keyword_apikeyMY_KEY/str str name=concept_apikeyMY_KEY/str str name=lang_apikeyMY_KEY/str str name=cat_apikeyMY_KEY/str str name=entities_apikeyMY_KEY/str str name=oc_licenseIDMY_SECOND_KEY/str /lst str name=analysisEngine/org/apache/uima/desc/OverridingParamsExtServicesAE.xml/str bool name=ignoreErrorsfalse/bool str name=logFieldtitle/str lst name=analyzeFields bool name=mergefalse/bool * arr name=fields* *strtitle/str* * /arr* /lst lst name=fieldMappings lst name=type str name=nameorg.apache.uima.alchemy.ts.concept.ConceptFS/str lst name=mapping *str name=featuretitle/str* * str name=fieldconcept/str* /lst /lst lst name=type str name=nameorg.apache.uima.SentenceAnnotation/str lst name=mapping * str name=featuretitle/str* * str name=fieldsentence/str* /lst /lst /lst /lst /processor and requestHandler name=/update class=solr.XmlUpdateRequestHandler lst name=defaults str name=update.processoruima/str /lst /requestHandler I am trying to index a simple document which looks like the following add doc field name=docid1456780001/field field name=*title*Canon powershow camera 9000/field /doc /add I am using curl to post this document on the /update end point and I am getting the following error *org.apache.solr.common.SolrException: processing error: null.* title=Canon powershow camera 9000, text=quot;Canon powershow camera 9000...quot; at org.apache.solr.uima.processor.UIMAUpdateRequestProcessor.processAdd(UIMAUpdateRequestProcessor.java:107) at org.apache.solr.handler.XMLLoader.processUpdate(XMLLoader.java:147) at org.apache.solr.handler.XMLLoader.load(XMLLoader.java:77) at org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:67) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:129) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1368) at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:356) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:252) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:243) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:210) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:224) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:175) at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:472) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:164) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:100) at org.apache.catalina.valves.AccessLogValve.invoke(AccessLogValve.java:929) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:118) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:405) at org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:279) at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:515) at org.apache.tomcat.util.net.JIoEndpoint$SocketProcessor.run(JIoEndpoint.java:300) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:680) Caused by: org.apache.solr.uima.processor.exception.FieldMappingException *at org.apache.solr.uima.processor.UIMAToSolrMapper.map(UIMAToSolrMapper.java:83) * * at org.apache.solr.uima.processor.UIMAUpdateRequestProcessor.processAdd(UIMAUpdateRequestProcessor.java:85) * * ... 23 more* What could be the problem? Thanks for your time
Re: How to connect Solr with external Zookeeper ensemble
What have you tried already? In particular, have you looked at http://wiki.apache.org/solr/SolrCloud Best Erick On Tue, Aug 16, 2011 at 2:22 PM, Sharath Jagannath shotsonclo...@gmail.com wrote: Hey, How could I connect my solr server with external zookeeper? Thanks, Sharath
Re: exceeded limit of maxWarmingSearchers ERROR
Naveen: See below: *NRT with Apache Solr 3.3 and RankingAlgorithm does need a commit for a document to become searchable*. Any document that you add through update becomes immediately searchable. So no need to commit from within your update client code. Since there is no commit, the cache does not have to be cleared or the old searchers closed or new searchers opened, and warmed (error that you are facing). Looking at the link which you mentioned is clearly what we wanted. But the real thing is that you have RA does need a commit for a document to become searchable (please take a look at bold sentence) . Yes, as said earlier you do not need a commit. A document becomes searchable as soon as you add it. Below is an example of adding a document with curl (this from the wiki at http://solr-ra.tgels.com/wiki/en/Near_Real_Time_Search_ver_3.x): curl http://localhost:8983/solr/update/csv?stream.file=/tmp/x1.csvencapsulator=%1f; There is no commit included. The contents of the document become immediately searchable. In future, for more loads, can it cater to Master Slave (Replication) and etc to scale and perform better? If yes, we would like to go for NRT and looking at the performance described in the article is acceptable. We were expecting the same real time performance for a single user. There are no changes to Master/Slave (replication) process. So any changes you have currently will work as before or if you enable replication later, it should still work as without NRT. What about multiple users, should we wait for 1-2 secs before calling the curl request to make SOLR perform better. Or internally it will handle with multiple request (multithreaded and etc). Again for updating documents, you do not have to change your current process or code. Everything remains the same, except that if you were including commit, you do not include commit in your update statements. There is no change to the existing update process so internally it will not queue or multi-thread updates. It is as in existing Solr functionality, there no changes to the existing setup. Regarding perform better, in the Wiki paper every update through curl adds (streams) 500 documents. So you could take this approach. (this was something that I chose randomly to test the performance but seems to be good) What would be doc size (10,000 docs) to allow JVM perform better? Have you done any kind of benchmarking in terms of multi threaded and multi user for NRT and also JVM tuning in terms of SOLR sever performance. Any kind of performance analysis would help us to decide quickly to switch over to NRT. The performance discussed in the wiki paper uses the MBArtists index. The MBArtists index is the index used as one of the examples in the book, Solr 1.4 Enterprise Search Server. You can download and build this index if you have the book or can also download the contents from musicbrainz.org. Each doc maybe about 100 bytes and has about 7 fields. Performance with wikipedia's xml dump, commenting out skipdoc field (include redirects) in the dataconfig.xml [ dataimport handler ], the update performance is about 15000 docs / sec (100 million docs), with the skipdoc enabled (does not skip redirects), the performance is about 1350 docs / sec [ time spent mostly converting validating/xml than actual update ] (about 11 million docs ). Documents in wikipedia can be quite big, at least avg size of about 2500-5000 bytes or more. I would suggest that you download and give NRT with Apache Solr 3.3 and RankingAlgorithm a try and get a feel of it as this would be the best way to see how your config works with it. Questions in terms for switching over to NRT, 1.Should we upgrade to SOLR 4.x ? 2. Any benchmarking (10,000 docs/secs). The question here is more specific the detail of individual doc (fields, number of fields, fields size, parameters affecting performance with faceting or w/o faceting) Please see the MBArtists index as discussed above. 3. What about multiple users ? A user in real time might be having an large doc size of .1 million. How to break and analyze which one is better (though it is our task to do). But still any kind of break up will help us. Imagine a user inbox. You maybe able to stream the documents in a set as in the example in the wiki. The example streams 500 documents at a time. The wiki paper has an example of a document that was used. You could copy/paste that to try it out. 4. JVM tuning and performance result based on Multithreaded environment. 5. Machine Details (RAM, CPU, and settings from SOLR perspective). Default Solr settings with the shipped jetty container. The startup script used is available when you download Solr 3.3 with RankingAlgorithm. It has mx set to 2Gb and uses the default collector with parallel collection enabled for the young generation. The system is a x86_64 Linux (2.6 kernel), 2 core (2.5Ghz) and uses internal
Re: Product data schema question
Why don't you use fields for each size? You can update our inventory only in the event of a size becoming avaiable or unavaviable. That would remove a lot of the load in inventory update. Anothe way is to treat each sku/inventory pair as a document. 2011/8/16 Jaeger, Jay - DOT jay.jae...@dot.wi.gov Not particularly. Just trying to do my part to answer some questions on the list. -Original Message- From: Steve Cerny [mailto:sjce...@gmail.com] Sent: Tuesday, August 16, 2011 11:49 AM To: solr-user@lucene.apache.org Subject: Re: Product data schema question Thanks Jay, if we come to a reasonable solution are you interested in the details? On Tue, Aug 16, 2011 at 11:44 AM, Jaeger, Jay - DOT jay.jae...@dot.wi.govwrote: No, I don't think so. A given core can only use one configuration and therefore only one schema, as far as I know, and a schema can only have one key. You could use two cores with two configurations (but that presumably wouldn't be much help). Solr is not a DBMS. It is an index. -Original Message- From: Steve Cerny [mailto:sjce...@gmail.com] Sent: Tuesday, August 16, 2011 11:37 AM To: solr-user@lucene.apache.org Subject: Re: Product data schema question Jay, this is great information. I don't know enough about Solr whether this is possible...Can we setup two indexes in the same core, one for product_catalog and the other for inventory? Then using a Solr query we could join the indexed content together. In Sql it would look like this select p.brand , p.model_name , p.sku , p.color_name , i.variant_id , i.variant_count from product_catalog p join inventory i on (p.sku = i.sku) On Tue, Aug 16, 2011 at 8:00 AM, Jaeger, Jay - DOT jay.jae...@dot.wi.gov wrote: On the surface, you could simply add some more fields to your schema. But as far as I can tell, you would have to have a separate Solr document for each SKU/size combination, and store the rest of the information (brand, model, color, SKU) redundantly and make the unique key a combination of the SKU and the size (presumably by having an additional field called sku_size - as far as I can tell Solr can't make a key up out of multiple fields). But, perhaps you shouldn't store that dynamic inventory information in Solr. Instead store a key that gets you to an inventory database organized by SKU and size. In such a schema, Size could be added as a multi-valued field to your existing schema, so you would know what sizes existed, so you could at least facet on that, maybe. That way Solr doesn't have to be updated every time the inventory changes. Of course, that won't help with the faceting on variant inventory. -Original Message- From: Steve Cerny [mailto:sjce...@gmail.com] Sent: Monday, August 15, 2011 6:29 PM To: solr-user@lucene.apache.org Subject: Product data schema question I'm working on an online eCommerce project and am having difficulties building the core / index schema. Here is the way we organize our product information in a normalized database. A product model has many SKUs (called colorways) A SKU has many sizes (called variants) A SKU size has associated inventory (called variant inventory) When we setup our product core we have the following field information Doc * brand * model name * SKU * color name Sample records are as follows * Haynes, Undershirt, 1234, white * Haynes, Undershirt, 1235, grey * Fruit of the Loom, Undershirt, 1236, white * Fruit of the Loom, Underwear, 1237, grey The issue I'm having is I want to add inventory to each size of each SKU for faceting. Example, SKU 1234 has sizes small, medium, large. Size small has 5 in stock, size medium 10, and size large 25. In a normalized data table I would have a separate table just for inventory and related it back to the SKU with a foreign key. How do I store size and inventory information effectively with Solr? -- Steve -- Steve -- Steve Cerny sjce...@gmail.com 715-302-0639 -- Alexander Ramos Jardim