The command you are using is not pointing to the specific solr index you created. The http://localhost:8983/solr needs to be changed to the URL for the core created. It should look like http://localhost:8983/solr/#/new_core_name.
On Tue, Apr 7, 2015 at 2:33 AM, Anchit Jain <[email protected]> wrote: > I followed instructions as given on your blog and created a new core for > nutch data and copied schema.xml of nutch into it. > Then I run the following command in nutch working directory > bin/nutch solrindex http://localhost:8983/solr crawl/crawldb/ -linkdb > crawl/linkdb/ crawl/segments/20150406231502/ -filter -normalize > > But then also the same error is coming as like previous runs. > > On Tue, 7 Apr 2015 at 10:00 Jeff Cocking <[email protected]> wrote: > > > Solr5 is multicore by default. You have not finished the install by > > setting up solr5's core. I would suggest you look at the link I sent to > > finish up your setup. > > > > After you finish your install your solr URL will be > > http://localhost:8983/solr/#/core_name. > > > > Jeff Cocking > > > > I apologize for my brevity. > > This was sent from my mobile device while I should be focusing on > > something else..... > > Like a meeting, driving, family, etc. > > > > > On Apr 6, 2015, at 11:16 PM, Anchit Jain <[email protected]> > > wrote: > > > > > > I have already installed Solr.I want to integrate it with nutch. > > > Whenever I try to issue this command to nutch > > > ""bin/nutch solrindex http://localhost:8983/solr crawl/crawldb/ > -linkdb > > > crawl/linkdb/ crawl/segments/20150406231502/ -filter -normalize" > > > > > > I always get a error > > > > > > Indexer: java.io.IOException: Job failed! > > > at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1357) > > > at org.apache.nutch.indexer.IndexingJob.index(IndexingJob.java:114) > > > at org.apache.nutch.indexer.IndexingJob.run(IndexingJob.java:176) > > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) > > > at org.apache.nutch.indexer.IndexingJob.main(IndexingJob.java:186) > > > > > > > > > > > > Here is the complete hadoop log for the process.I have underlined the > > error > > > part in it. > > > > > > 2015-04-07 09:38:06,613 INFO indexer.IndexingJob - Indexer: starting > at > > > 2015-04-07 09:38:06 > > > 2015-04-07 09:38:06,684 INFO indexer.IndexingJob - Indexer: deleting > > gone > > > documents: false > > > 2015-04-07 09:38:06,685 INFO indexer.IndexingJob - Indexer: URL > > filtering: > > > true > > > 2015-04-07 09:38:06,685 INFO indexer.IndexingJob - Indexer: URL > > > normalizing: true > > > 2015-04-07 09:38:06,893 INFO indexer.IndexWriters - Adding > > > org.apache.nutch.indexwriter.solr.SolrIndexWriter > > > 2015-04-07 09:38:06,893 INFO indexer.IndexingJob - Active > IndexWriters : > > > SOLRIndexWriter > > > solr.server.url : URL of the SOLR instance (mandatory) > > > solr.commit.size : buffer size when sending to SOLR (default 1000) > > > solr.mapping.file : name of the mapping file for fields (default > > > solrindex-mapping.xml) > > > solr.auth : use authentication (default false) > > > solr.auth.username : use authentication (default false) > > > solr.auth : username for authentication > > > solr.auth.password : password for authentication > > > > > > > > > 2015-04-07 09:38:06,898 INFO indexer.IndexerMapReduce - > > IndexerMapReduce: > > > crawldb: crawl/crawldb > > > 2015-04-07 09:38:06,898 INFO indexer.IndexerMapReduce - > > IndexerMapReduce: > > > linkdb: crawl/linkdb > > > 2015-04-07 09:38:06,898 INFO indexer.IndexerMapReduce - > > IndexerMapReduces: > > > adding segment: crawl/segments/20150406231502 > > > 2015-04-07 09:38:07,036 WARN util.NativeCodeLoader - Unable to load > > > native-hadoop library for your platform... using builtin-java classes > > where > > > applicable > > > 2015-04-07 09:38:07,540 INFO anchor.AnchorIndexingFilter - Anchor > > > deduplication is: off > > > 2015-04-07 09:38:07,565 INFO regex.RegexURLNormalizer - can't find > rules > > > for scope 'indexer', using default > > > 2015-04-07 09:38:09,552 INFO regex.RegexURLNormalizer - can't find > rules > > > for scope 'indexer', using default > > > 2015-04-07 09:38:10,642 INFO regex.RegexURLNormalizer - can't find > rules > > > for scope 'indexer', using default > > > 2015-04-07 09:38:10,734 INFO regex.RegexURLNormalizer - can't find > rules > > > for scope 'indexer', using default > > > 2015-04-07 09:38:10,895 INFO regex.RegexURLNormalizer - can't find > rules > > > for scope 'indexer', using default > > > 2015-04-07 09:38:11,088 INFO regex.RegexURLNormalizer - can't find > rules > > > for scope 'indexer', using default > > > 2015-04-07 09:38:11,219 INFO indexer.IndexWriters - Adding > > > org.apache.nutch.indexwriter.solr.SolrIndexWriter > > > 2015-04-07 09:38:11,237 INFO solr.SolrMappingReader - source: content > > > dest: content > > > 2015-04-07 09:38:11,237 INFO solr.SolrMappingReader - source: title > > dest: > > > title > > > 2015-04-07 09:38:11,237 INFO solr.SolrMappingReader - source: host > dest: > > > host > > > 2015-04-07 09:38:11,237 INFO solr.SolrMappingReader - source: segment > > > dest: segment > > > 2015-04-07 09:38:11,237 INFO solr.SolrMappingReader - source: boost > > dest: > > > boost > > > 2015-04-07 09:38:11,237 INFO solr.SolrMappingReader - source: digest > > dest: > > > digest > > > 2015-04-07 09:38:11,237 INFO solr.SolrMappingReader - source: tstamp > > dest: > > > tstamp > > > 2015-04-07 09:38:11,526 INFO solr.SolrIndexWriter - Indexing 250 > > documents > > > 2015-04-07 09:38:11,526 INFO solr.SolrIndexWriter - Deleting 0 > documents > > > 2015-04-07 09:38:11,644 INFO solr.SolrIndexWriter - Indexing 250 > > documents > > > *2015-04-07 09:38:11,699 WARN mapred.LocalJobRunner - > > > job_local1245074757_0001* > > > *org.apache.solr.common.SolrException: Not Found* > > > > > > *Not Found* > > > > > > *request: http://localhost:8983/solr/update?wt=javabin&version=2 > > > <http://localhost:8983/solr/update?wt=javabin&version=2>* > > > * at > > > org.apache.solr.client.solrj.impl.CommonsHttpSolrServer. > > request(CommonsHttpSolrServer.java:430)* > > > * at > > > org.apache.solr.client.solrj.impl.CommonsHttpSolrServer. > > request(CommonsHttpSolrServer.java:244)* > > > * at > > > org.apache.solr.client.solrj.request.AbstractUpdateRequest. > > process(AbstractUpdateRequest.java:105)* > > > * at > > > org.apache.nutch.indexwriter.solr.SolrIndexWriter.write( > > SolrIndexWriter.java:135)* > > > * at org.apache.nutch.indexer.IndexWriters.write(IndexWriters.java:88)* > > > * at > > > org.apache.nutch.indexer.IndexerOutputFormat$1.write( > > IndexerOutputFormat.java:50)* > > > * at > > > org.apache.nutch.indexer.IndexerOutputFormat$1.write( > > IndexerOutputFormat.java:41)* > > > * at > > > org.apache.hadoop.mapred.ReduceTask$OldTrackingRecordWriter.write( > > ReduceTask.java:458)* > > > * at > org.apache.hadoop.mapred.ReduceTask$3.collect(ReduceTask.java:500)* > > > * at > > > org.apache.nutch.indexer.IndexerMapReduce.reduce( > > IndexerMapReduce.java:323)* > > > * at > > > org.apache.nutch.indexer.IndexerMapReduce.reduce( > > IndexerMapReduce.java:53)* > > > * at org.apache.hadoop.mapred.ReduceTask.runOldReducer( > > ReduceTask.java:522)* > > > * at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:421)* > > > * at > > > org.apache.hadoop.mapred.LocalJobRunner$Job.run( > > LocalJobRunner.java:398)* > > > *2015-04-07 09:38:12,408 ERROR indexer.IndexingJob - Indexer: > > > java.io.IOException: Job failed!* > > > * at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1357)* > > > * at org.apache.nutch.indexer.IndexingJob.index(IndexingJob.java:114)* > > > * at org.apache.nutch.indexer.IndexingJob.run(IndexingJob.java:176)* > > > * at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)* > > > * at org.apache.nutch.indexer.IndexingJob.main(IndexingJob.java:186)* > > >> On Tue, 7 Apr 2015 at 03:18 Jeff Cocking <[email protected]> > > wrote: > > >> > > >> With Solr5.0.0 you can skip that step. Solr will auto create your > > schema > > >> document based on the data being provided. > > >> > > >> One of the new features with Solr5 is the install/service feature. I > > did a > > >> quick write up on how to install Solr5 on Centos. Might be something > > >> useful there for you. > > >> > > >> http://www.cocking.com/apache-solr-5-0-install-on-centos-7/ > > >> > > >> jeff > > >> > > >> On Mon, Apr 6, 2015 at 3:13 PM, Anchit Jain <[email protected] > > > > >> wrote: > > >> > > >>> I want to index nutch results using *Solr 5.0* but as mentioned in > > >>> https://wiki.apache.org/nutch/NutchTutorial there is no directory > > >>> ${APACHE_SOLR_HOME}/example/solr/collection1/conf/ > > >>> in solr 5.0 . So where I have to copy *schema.xml*? > > >>> Also there is no *start.jar* present in example directory. > > >> > > >

