There can be numerous reasons....Hosts.conf, firewall, etc. These are all unique to your system.
Have you viewed the solr admin panel via a browser? This is a critical step in the installation. This validates SOLR can accept HTTP commands. On Tue, Apr 7, 2015 at 9:53 AM, Anchit Jain <[email protected]> wrote: > I created a new core named *foo*. Than I copied the *schema.xml* from > *nutch* into *var/solr/data/foo/conf* with changes as described in* > https://wiki.apache.org/nutch/NutchTutorial*. > I changed the url to*http://localhost:8983/solr/#/foo* > so new command is > "*bin/nutch solrindex http://localhost:8983/solr/#/foo crawl/crawldb/ > -linkdb crawl/linkdb/ crawl/segments/20150406231502/ -filter -normalize*" > But now I am getting error > *org.apache.solr.common.SolrException: HTTP method POST is not supported > by this URL* > * > * > Is some other change is also required in URL to support POST requests? > > Full log > > 2015-04-07 20:10:56,068 INFO indexer.IndexingJob - Indexer: starting at > 2015-04-07 20:10:56 > 2015-04-07 20:10:56,178 INFO indexer.IndexingJob - Indexer: deleting gone > documents: false > 2015-04-07 20:10:56,178 INFO indexer.IndexingJob - Indexer: URL > filtering: true > 2015-04-07 20:10:56,178 INFO indexer.IndexingJob - Indexer: URL > normalizing: true > 2015-04-07 20:10:56,727 INFO indexer.IndexWriters - Adding > org.apache.nutch.indexwriter.solr.SolrIndexWriter > 2015-04-07 20:10:56,727 INFO indexer.IndexingJob - Active IndexWriters : > SOLRIndexWriter > solr.server.url : URL of the SOLR instance (mandatory) > solr.commit.size : buffer size when sending to SOLR (default 1000) > solr.mapping.file : name of the mapping file for fields (default > solrindex-mapping.xml) > solr.auth : use authentication (default false) > solr.auth.username : use authentication (default false) > solr.auth : username for authentication > solr.auth.password : password for authentication > > > 2015-04-07 20:10:56,772 INFO indexer.IndexerMapReduce - IndexerMapReduce: > crawldb: crawl/crawldb > 2015-04-07 20:10:56,772 INFO indexer.IndexerMapReduce - IndexerMapReduce: > linkdb: crawl/linkdb > 2015-04-07 20:10:56,772 INFO indexer.IndexerMapReduce - > IndexerMapReduces: adding segment: crawl/segments/20150406231502 > 2015-04-07 20:10:57,205 WARN util.NativeCodeLoader - Unable to load > native-hadoop library for your platform... using builtin-java classes where > applicable > 2015-04-07 20:10:58,020 INFO anchor.AnchorIndexingFilter - Anchor > deduplication is: off > 2015-04-07 20:10:58,134 INFO regex.RegexURLNormalizer - can't find rules > for scope 'indexer', using default > 2015-04-07 20:11:00,114 INFO regex.RegexURLNormalizer - can't find rules > for scope 'indexer', using default > 2015-04-07 20:11:01,205 INFO regex.RegexURLNormalizer - can't find rules > for scope 'indexer', using default > 2015-04-07 20:11:01,344 INFO regex.RegexURLNormalizer - can't find rules > for scope 'indexer', using default > 2015-04-07 20:11:01,577 INFO regex.RegexURLNormalizer - can't find rules > for scope 'indexer', using default > 2015-04-07 20:11:01,788 INFO regex.RegexURLNormalizer - can't find rules > for scope 'indexer', using default > 2015-04-07 20:11:01,921 INFO indexer.IndexWriters - Adding > org.apache.nutch.indexwriter.solr.SolrIndexWriter > 2015-04-07 20:11:01,986 INFO solr.SolrMappingReader - source: content > dest: content > 2015-04-07 20:11:01,986 INFO solr.SolrMappingReader - source: title dest: > title > 2015-04-07 20:11:01,986 INFO solr.SolrMappingReader - source: host dest: > host > 2015-04-07 20:11:01,986 INFO solr.SolrMappingReader - source: segment > dest: segment > 2015-04-07 20:11:01,987 INFO solr.SolrMappingReader - source: boost dest: > boost > 2015-04-07 20:11:01,987 INFO solr.SolrMappingReader - source: digest > dest: digest > 2015-04-07 20:11:01,987 INFO solr.SolrMappingReader - source: tstamp > dest: tstamp > 2015-04-07 20:11:02,266 INFO solr.SolrIndexWriter - Indexing 250 documents > 2015-04-07 20:11:02,267 INFO solr.SolrIndexWriter - Deleting 0 documents > 2015-04-07 20:11:02,512 INFO solr.SolrIndexWriter - Indexing 250 documents > *2015-04-07 20:11:02,576 WARN mapred.LocalJobRunner - > job_local1831338118_0001* > *org.apache.solr.common.SolrException: HTTP method POST is not supported > by this URL* > * > * > *HTTP method POST is not supported by this URL* > > request: http://localhost:8983/solr/ > at org.apache.solr.client.solrj.impl.CommonsHttpSolrServer. > request(CommonsHttpSolrServer.java:430) > at org.apache.solr.client.solrj.impl.CommonsHttpSolrServer. > request(CommonsHttpSolrServer.java:244) > at org.apache.solr.client.solrj.request.AbstractUpdateRequest. > process(AbstractUpdateRequest.java:105) > at org.apache.nutch.indexwriter.solr.SolrIndexWriter.write( > SolrIndexWriter.java:135) > at org.apache.nutch.indexer.IndexWriters.write(IndexWriters.java:88) > at org.apache.nutch.indexer.IndexerOutputFormat$1.write( > IndexerOutputFormat.java:50) > at org.apache.nutch.indexer.IndexerOutputFormat$1.write( > IndexerOutputFormat.java:41) > at org.apache.hadoop.mapred.ReduceTask$OldTrackingRecordWriter.write( > ReduceTask.java:458) > at org.apache.hadoop.mapred.ReduceTask$3.collect(ReduceTask.java:500) > at org.apache.nutch.indexer.IndexerMapReduce.reduce( > IndexerMapReduce.java:323) > at org.apache.nutch.indexer.IndexerMapReduce.reduce( > IndexerMapReduce.java:53) > at org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:522) > at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:421) > at org.apache.hadoop.mapred.LocalJobRunner$Job.run( > LocalJobRunner.java:398) > 2015-04-07 20:11:02,724 ERROR indexer.IndexingJob - Indexer: > java.io.IOException: Job failed! > at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1357) > at org.apache.nutch.indexer.IndexingJob.index(IndexingJob.java:114) > at org.apache.nutch.indexer.IndexingJob.run(IndexingJob.java:176) > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) > at org.apache.nutch.indexer.IndexingJob.main(IndexingJob.java:186) > > > On Tuesday 07 April 2015 06:53 PM, Jeff Cocking wrote: > >> The command you are using is not pointing to the specific solr index you >> created. The http://localhost:8983/solr needs to be changed to the URL >> for >> the core created. It should look like >> http://localhost:8983/solr/#/new_core_name. >> >> >> On Tue, Apr 7, 2015 at 2:33 AM, Anchit Jain <[email protected]> >> wrote: >> >> I followed instructions as given on your blog and created a new core for >>> nutch data and copied schema.xml of nutch into it. >>> Then I run the following command in nutch working directory >>> bin/nutch solrindex http://localhost:8983/solr crawl/crawldb/ -linkdb >>> crawl/linkdb/ crawl/segments/20150406231502/ -filter -normalize >>> >>> But then also the same error is coming as like previous runs. >>> >>> On Tue, 7 Apr 2015 at 10:00 Jeff Cocking <[email protected]> wrote: >>> >>> Solr5 is multicore by default. You have not finished the install by >>>> setting up solr5's core. I would suggest you look at the link I sent to >>>> finish up your setup. >>>> >>>> After you finish your install your solr URL will be >>>> http://localhost:8983/solr/#/core_name. >>>> >>>> Jeff Cocking >>>> >>>> I apologize for my brevity. >>>> This was sent from my mobile device while I should be focusing on >>>> something else..... >>>> Like a meeting, driving, family, etc. >>>> >>>> On Apr 6, 2015, at 11:16 PM, Anchit Jain <[email protected]> >>>>> >>>> wrote: >>>> >>>>> I have already installed Solr.I want to integrate it with nutch. >>>>> Whenever I try to issue this command to nutch >>>>> ""bin/nutch solrindex http://localhost:8983/solr crawl/crawldb/ >>>>> >>>> -linkdb >>> >>>> crawl/linkdb/ crawl/segments/20150406231502/ -filter -normalize" >>>>> >>>>> I always get a error >>>>> >>>>> Indexer: java.io.IOException: Job failed! >>>>> at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1357) >>>>> at org.apache.nutch.indexer.IndexingJob.index(IndexingJob.java:114) >>>>> at org.apache.nutch.indexer.IndexingJob.run(IndexingJob.java:176) >>>>> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) >>>>> at org.apache.nutch.indexer.IndexingJob.main(IndexingJob.java:186) >>>>> >>>>> >>>>> >>>>> Here is the complete hadoop log for the process.I have underlined the >>>>> >>>> error >>>> >>>>> part in it. >>>>> >>>>> 2015-04-07 09:38:06,613 INFO indexer.IndexingJob - Indexer: starting >>>>> >>>> at >>> >>>> 2015-04-07 09:38:06 >>>>> 2015-04-07 09:38:06,684 INFO indexer.IndexingJob - Indexer: deleting >>>>> >>>> gone >>>> >>>>> documents: false >>>>> 2015-04-07 09:38:06,685 INFO indexer.IndexingJob - Indexer: URL >>>>> >>>> filtering: >>>> >>>>> true >>>>> 2015-04-07 09:38:06,685 INFO indexer.IndexingJob - Indexer: URL >>>>> normalizing: true >>>>> 2015-04-07 09:38:06,893 INFO indexer.IndexWriters - Adding >>>>> org.apache.nutch.indexwriter.solr.SolrIndexWriter >>>>> 2015-04-07 09:38:06,893 INFO indexer.IndexingJob - Active >>>>> >>>> IndexWriters : >>> >>>> SOLRIndexWriter >>>>> solr.server.url : URL of the SOLR instance (mandatory) >>>>> solr.commit.size : buffer size when sending to SOLR (default 1000) >>>>> solr.mapping.file : name of the mapping file for fields (default >>>>> solrindex-mapping.xml) >>>>> solr.auth : use authentication (default false) >>>>> solr.auth.username : use authentication (default false) >>>>> solr.auth : username for authentication >>>>> solr.auth.password : password for authentication >>>>> >>>>> >>>>> 2015-04-07 09:38:06,898 INFO indexer.IndexerMapReduce - >>>>> >>>> IndexerMapReduce: >>>> >>>>> crawldb: crawl/crawldb >>>>> 2015-04-07 09:38:06,898 INFO indexer.IndexerMapReduce - >>>>> >>>> IndexerMapReduce: >>>> >>>>> linkdb: crawl/linkdb >>>>> 2015-04-07 09:38:06,898 INFO indexer.IndexerMapReduce - >>>>> >>>> IndexerMapReduces: >>>> >>>>> adding segment: crawl/segments/20150406231502 >>>>> 2015-04-07 09:38:07,036 WARN util.NativeCodeLoader - Unable to load >>>>> native-hadoop library for your platform... using builtin-java classes >>>>> >>>> where >>>> >>>>> applicable >>>>> 2015-04-07 09:38:07,540 INFO anchor.AnchorIndexingFilter - Anchor >>>>> deduplication is: off >>>>> 2015-04-07 09:38:07,565 INFO regex.RegexURLNormalizer - can't find >>>>> >>>> rules >>> >>>> for scope 'indexer', using default >>>>> 2015-04-07 09:38:09,552 INFO regex.RegexURLNormalizer - can't find >>>>> >>>> rules >>> >>>> for scope 'indexer', using default >>>>> 2015-04-07 09:38:10,642 INFO regex.RegexURLNormalizer - can't find >>>>> >>>> rules >>> >>>> for scope 'indexer', using default >>>>> 2015-04-07 09:38:10,734 INFO regex.RegexURLNormalizer - can't find >>>>> >>>> rules >>> >>>> for scope 'indexer', using default >>>>> 2015-04-07 09:38:10,895 INFO regex.RegexURLNormalizer - can't find >>>>> >>>> rules >>> >>>> for scope 'indexer', using default >>>>> 2015-04-07 09:38:11,088 INFO regex.RegexURLNormalizer - can't find >>>>> >>>> rules >>> >>>> for scope 'indexer', using default >>>>> 2015-04-07 09:38:11,219 INFO indexer.IndexWriters - Adding >>>>> org.apache.nutch.indexwriter.solr.SolrIndexWriter >>>>> 2015-04-07 09:38:11,237 INFO solr.SolrMappingReader - source: content >>>>> dest: content >>>>> 2015-04-07 09:38:11,237 INFO solr.SolrMappingReader - source: title >>>>> >>>> dest: >>>> >>>>> title >>>>> 2015-04-07 09:38:11,237 INFO solr.SolrMappingReader - source: host >>>>> >>>> dest: >>> >>>> host >>>>> 2015-04-07 09:38:11,237 INFO solr.SolrMappingReader - source: segment >>>>> dest: segment >>>>> 2015-04-07 09:38:11,237 INFO solr.SolrMappingReader - source: boost >>>>> >>>> dest: >>>> >>>>> boost >>>>> 2015-04-07 09:38:11,237 INFO solr.SolrMappingReader - source: digest >>>>> >>>> dest: >>>> >>>>> digest >>>>> 2015-04-07 09:38:11,237 INFO solr.SolrMappingReader - source: tstamp >>>>> >>>> dest: >>>> >>>>> tstamp >>>>> 2015-04-07 09:38:11,526 INFO solr.SolrIndexWriter - Indexing 250 >>>>> >>>> documents >>>> >>>>> 2015-04-07 09:38:11,526 INFO solr.SolrIndexWriter - Deleting 0 >>>>> >>>> documents >>> >>>> 2015-04-07 09:38:11,644 INFO solr.SolrIndexWriter - Indexing 250 >>>>> >>>> documents >>>> >>>>> *2015-04-07 09:38:11,699 WARN mapred.LocalJobRunner - >>>>> job_local1245074757_0001* >>>>> *org.apache.solr.common.SolrException: Not Found* >>>>> >>>>> *Not Found* >>>>> >>>>> *request: http://localhost:8983/solr/update?wt=javabin&version=2 >>>>> <http://localhost:8983/solr/update?wt=javabin&version=2>* >>>>> * at >>>>> org.apache.solr.client.solrj.impl.CommonsHttpSolrServer. >>>>> >>>> request(CommonsHttpSolrServer.java:430)* >>>> >>>>> * at >>>>> org.apache.solr.client.solrj.impl.CommonsHttpSolrServer. >>>>> >>>> request(CommonsHttpSolrServer.java:244)* >>>> >>>>> * at >>>>> org.apache.solr.client.solrj.request.AbstractUpdateRequest. >>>>> >>>> process(AbstractUpdateRequest.java:105)* >>>> >>>>> * at >>>>> org.apache.nutch.indexwriter.solr.SolrIndexWriter.write( >>>>> >>>> SolrIndexWriter.java:135)* >>>> >>>>> * at org.apache.nutch.indexer.IndexWriters.write( >>>>> IndexWriters.java:88)* >>>>> * at >>>>> org.apache.nutch.indexer.IndexerOutputFormat$1.write( >>>>> >>>> IndexerOutputFormat.java:50)* >>>> >>>>> * at >>>>> org.apache.nutch.indexer.IndexerOutputFormat$1.write( >>>>> >>>> IndexerOutputFormat.java:41)* >>>> >>>>> * at >>>>> org.apache.hadoop.mapred.ReduceTask$OldTrackingRecordWriter.write( >>>>> >>>> ReduceTask.java:458)* >>>> >>>>> * at >>>>> >>>> org.apache.hadoop.mapred.ReduceTask$3.collect(ReduceTask.java:500)* >>> >>>> * at >>>>> org.apache.nutch.indexer.IndexerMapReduce.reduce( >>>>> >>>> IndexerMapReduce.java:323)* >>>> >>>>> * at >>>>> org.apache.nutch.indexer.IndexerMapReduce.reduce( >>>>> >>>> IndexerMapReduce.java:53)* >>>> >>>>> * at org.apache.hadoop.mapred.ReduceTask.runOldReducer( >>>>> >>>> ReduceTask.java:522)* >>>> >>>>> * at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:421)* >>>>> * at >>>>> org.apache.hadoop.mapred.LocalJobRunner$Job.run( >>>>> >>>> LocalJobRunner.java:398)* >>>> >>>>> *2015-04-07 09:38:12,408 ERROR indexer.IndexingJob - Indexer: >>>>> java.io.IOException: Job failed!* >>>>> * at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1357)* >>>>> * at org.apache.nutch.indexer.IndexingJob.index(IndexingJob.java:114)* >>>>> * at org.apache.nutch.indexer.IndexingJob.run(IndexingJob.java:176)* >>>>> * at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)* >>>>> * at org.apache.nutch.indexer.IndexingJob.main(IndexingJob.java:186)* >>>>> >>>>>> On Tue, 7 Apr 2015 at 03:18 Jeff Cocking <[email protected]> >>>>>> >>>>> wrote: >>>> >>>>> With Solr5.0.0 you can skip that step. Solr will auto create your >>>>>> >>>>> schema >>>> >>>>> document based on the data being provided. >>>>>> >>>>>> One of the new features with Solr5 is the install/service feature. I >>>>>> >>>>> did a >>>> >>>>> quick write up on how to install Solr5 on Centos. Might be something >>>>>> useful there for you. >>>>>> >>>>>> http://www.cocking.com/apache-solr-5-0-install-on-centos-7/ >>>>>> >>>>>> jeff >>>>>> >>>>>> On Mon, Apr 6, 2015 at 3:13 PM, Anchit Jain <[email protected] >>>>>> wrote: >>>>>> >>>>>> I want to index nutch results using *Solr 5.0* but as mentioned in >>>>>>> https://wiki.apache.org/nutch/NutchTutorial there is no directory >>>>>>> ${APACHE_SOLR_HOME}/example/solr/collection1/conf/ >>>>>>> in solr 5.0 . So where I have to copy *schema.xml*? >>>>>>> Also there is no *start.jar* present in example directory. >>>>>>> >>>>>> >

