I don't want to sound negative, but I think it is a valid question to consider - for the lack of information and certain mental rigidity may make it sound bad - first of all, it is probably not for few gigabytes of data and I can imagine that building indexes at the side when data lives is much faster/cheaper, then sending data to solr - if we think the index is the product of the map, then the 'reduce' part may be this http://wiki.apache.org/solr/MergingSolrIndexes
I don't really know enough about CloudSolrServer and how to fit the cloud there roman On Fri, Jul 5, 2013 at 12:23 PM, Jack Krupansky <j...@basetechnology.com>wrote: > Software developers are sometimes compensated based on the degree of > complexity that they deal with. > > And managers are sometimes compensated based on the number of people they > manage, as well as the degree of complexity of what they manage. > > And... training organizations can charge more and have a larger pool of > eager customers when the subject matter has higher complexity. > > And... consultants and contractors will be in higher demand and able to > charge more, based on the degree of complexity that they have mastered. > > So, more complexity results in greater opportunity for higher income! > > (Oh, and, writers and book authors have more to write about and readers > are more eager to purchase those writings as well, especially if the > subject matter is constantly changing.) > > Somebody please remind me I said this any time you catch me trying to > argue for Solr to be made simpler and easier to use! > > -- Jack Krupansky > > -----Original Message----- From: Walter Underwood > Sent: Friday, July 05, 2013 12:11 PM > To: solr-user@lucene.apache.org > Subject: Re: Sending Documents via SolrServer as MapReduce Jobs at Solrj > > > Why is it better to require another large software system (Hadoop), when > it works fine without it? > > That just sounds like more stuff to configure, misconfigure, and cause > problems with indexing. > > wunder > > On Jul 5, 2013, at 4:48 AM, Furkan KAMACI wrote: > > We are using Nutch to crawl web sites and it stores documents at Hbase. >> Nutch uses Solrj to send documents to be indexed. We have Hadoop at our >> ecosystem as well. I think that there should be an implementation at Solrj >> that sends documents (via CloudSolrServer or something like that) as >> MapReduce jobs. Is there any implentation for it or is it not a good idea? >> > > >