There can be numerous reasons....Hosts.conf, firewall, etc.  These are all
unique to your system.

Have you viewed the solr admin panel via a browser?  This is a critical
step in the installation.  This validates SOLR can accept HTTP commands.

On Tue, Apr 7, 2015 at 9:53 AM, Anchit Jain <[email protected]>
wrote:

> I created a new core named *foo*. Than I copied the *schema.xml* from
> *nutch* into *var/solr/data/foo/conf* with changes as described in*
> https://wiki.apache.org/nutch/NutchTutorial*.
> I changed the url to*http://localhost:8983/solr/#/foo*
> so new command is
> "*bin/nutch solrindex http://localhost:8983/solr/#/foo crawl/crawldb/
> -linkdb crawl/linkdb/ crawl/segments/20150406231502/ -filter -normalize*"
>  But now I am getting error
> *org.apache.solr.common.SolrException: HTTP method POST is not supported
> by this URL*
> *
> *
> Is some other change is also required in URL to support POST requests?
>
> Full log
>
> 2015-04-07 20:10:56,068 INFO  indexer.IndexingJob - Indexer: starting at
> 2015-04-07 20:10:56
> 2015-04-07 20:10:56,178 INFO  indexer.IndexingJob - Indexer: deleting gone
> documents: false
> 2015-04-07 20:10:56,178 INFO  indexer.IndexingJob - Indexer: URL
> filtering: true
> 2015-04-07 20:10:56,178 INFO  indexer.IndexingJob - Indexer: URL
> normalizing: true
> 2015-04-07 20:10:56,727 INFO  indexer.IndexWriters - Adding
> org.apache.nutch.indexwriter.solr.SolrIndexWriter
> 2015-04-07 20:10:56,727 INFO  indexer.IndexingJob - Active IndexWriters :
> SOLRIndexWriter
> solr.server.url : URL of the SOLR instance (mandatory)
> solr.commit.size : buffer size when sending to SOLR (default 1000)
> solr.mapping.file : name of the mapping file for fields (default
> solrindex-mapping.xml)
> solr.auth : use authentication (default false)
> solr.auth.username : use authentication (default false)
> solr.auth : username for authentication
> solr.auth.password : password for authentication
>
>
> 2015-04-07 20:10:56,772 INFO  indexer.IndexerMapReduce - IndexerMapReduce:
> crawldb: crawl/crawldb
> 2015-04-07 20:10:56,772 INFO  indexer.IndexerMapReduce - IndexerMapReduce:
> linkdb: crawl/linkdb
> 2015-04-07 20:10:56,772 INFO  indexer.IndexerMapReduce -
> IndexerMapReduces: adding segment: crawl/segments/20150406231502
> 2015-04-07 20:10:57,205 WARN  util.NativeCodeLoader - Unable to load
> native-hadoop library for your platform... using builtin-java classes where
> applicable
> 2015-04-07 20:10:58,020 INFO  anchor.AnchorIndexingFilter - Anchor
> deduplication is: off
> 2015-04-07 20:10:58,134 INFO  regex.RegexURLNormalizer - can't find rules
> for scope 'indexer', using default
> 2015-04-07 20:11:00,114 INFO  regex.RegexURLNormalizer - can't find rules
> for scope 'indexer', using default
> 2015-04-07 20:11:01,205 INFO  regex.RegexURLNormalizer - can't find rules
> for scope 'indexer', using default
> 2015-04-07 20:11:01,344 INFO  regex.RegexURLNormalizer - can't find rules
> for scope 'indexer', using default
> 2015-04-07 20:11:01,577 INFO  regex.RegexURLNormalizer - can't find rules
> for scope 'indexer', using default
> 2015-04-07 20:11:01,788 INFO  regex.RegexURLNormalizer - can't find rules
> for scope 'indexer', using default
> 2015-04-07 20:11:01,921 INFO  indexer.IndexWriters - Adding
> org.apache.nutch.indexwriter.solr.SolrIndexWriter
> 2015-04-07 20:11:01,986 INFO  solr.SolrMappingReader - source: content
> dest: content
> 2015-04-07 20:11:01,986 INFO  solr.SolrMappingReader - source: title dest:
> title
> 2015-04-07 20:11:01,986 INFO  solr.SolrMappingReader - source: host dest:
> host
> 2015-04-07 20:11:01,986 INFO  solr.SolrMappingReader - source: segment
> dest: segment
> 2015-04-07 20:11:01,987 INFO  solr.SolrMappingReader - source: boost dest:
> boost
> 2015-04-07 20:11:01,987 INFO  solr.SolrMappingReader - source: digest
> dest: digest
> 2015-04-07 20:11:01,987 INFO  solr.SolrMappingReader - source: tstamp
> dest: tstamp
> 2015-04-07 20:11:02,266 INFO  solr.SolrIndexWriter - Indexing 250 documents
> 2015-04-07 20:11:02,267 INFO  solr.SolrIndexWriter - Deleting 0 documents
> 2015-04-07 20:11:02,512 INFO  solr.SolrIndexWriter - Indexing 250 documents
> *2015-04-07 20:11:02,576 WARN  mapred.LocalJobRunner -
> job_local1831338118_0001*
> *org.apache.solr.common.SolrException: HTTP method POST is not supported
> by this URL*
> *
> *
> *HTTP method POST is not supported by this URL*
>
> request: http://localhost:8983/solr/
> at org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.
> request(CommonsHttpSolrServer.java:430)
> at org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.
> request(CommonsHttpSolrServer.java:244)
> at org.apache.solr.client.solrj.request.AbstractUpdateRequest.
> process(AbstractUpdateRequest.java:105)
> at org.apache.nutch.indexwriter.solr.SolrIndexWriter.write(
> SolrIndexWriter.java:135)
> at org.apache.nutch.indexer.IndexWriters.write(IndexWriters.java:88)
> at org.apache.nutch.indexer.IndexerOutputFormat$1.write(
> IndexerOutputFormat.java:50)
> at org.apache.nutch.indexer.IndexerOutputFormat$1.write(
> IndexerOutputFormat.java:41)
> at org.apache.hadoop.mapred.ReduceTask$OldTrackingRecordWriter.write(
> ReduceTask.java:458)
> at org.apache.hadoop.mapred.ReduceTask$3.collect(ReduceTask.java:500)
> at org.apache.nutch.indexer.IndexerMapReduce.reduce(
> IndexerMapReduce.java:323)
> at org.apache.nutch.indexer.IndexerMapReduce.reduce(
> IndexerMapReduce.java:53)
> at org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:522)
> at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:421)
> at org.apache.hadoop.mapred.LocalJobRunner$Job.run(
> LocalJobRunner.java:398)
> 2015-04-07 20:11:02,724 ERROR indexer.IndexingJob - Indexer:
> java.io.IOException: Job failed!
> at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1357)
> at org.apache.nutch.indexer.IndexingJob.index(IndexingJob.java:114)
> at org.apache.nutch.indexer.IndexingJob.run(IndexingJob.java:176)
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
> at org.apache.nutch.indexer.IndexingJob.main(IndexingJob.java:186)
>
>
> On Tuesday 07 April 2015 06:53 PM, Jeff Cocking wrote:
>
>> The command you are using is not pointing to the specific solr index you
>> created.  The http://localhost:8983/solr needs to be changed to the URL
>> for
>> the core created.  It should look like
>> http://localhost:8983/solr/#/new_core_name.
>>
>>
>> On Tue, Apr 7, 2015 at 2:33 AM, Anchit Jain <[email protected]>
>> wrote:
>>
>>  I followed instructions as given on your blog and created a new core for
>>> nutch data and copied schema.xml of nutch into it.
>>> Then I run the following command in nutch working directory
>>> bin/nutch solrindex http://localhost:8983/solr crawl/crawldb/ -linkdb
>>> crawl/linkdb/ crawl/segments/20150406231502/ -filter -normalize
>>>
>>> But then also the same error is coming as like previous runs.
>>>
>>> On Tue, 7 Apr 2015 at 10:00 Jeff Cocking <[email protected]> wrote:
>>>
>>>  Solr5 is multicore by default. You have not finished the install by
>>>> setting up solr5's core.  I would suggest you look at the link I sent to
>>>> finish up your setup.
>>>>
>>>> After you finish your install your solr URL will be
>>>> http://localhost:8983/solr/#/core_name.
>>>>
>>>> Jeff Cocking
>>>>
>>>> I apologize for my brevity.
>>>> This was sent from my mobile device while I should be focusing on
>>>> something else.....
>>>> Like a meeting, driving, family, etc.
>>>>
>>>>  On Apr 6, 2015, at 11:16 PM, Anchit Jain <[email protected]>
>>>>>
>>>> wrote:
>>>>
>>>>> I have already installed Solr.I want to integrate it with nutch.
>>>>> Whenever I try to issue this command to nutch
>>>>> ""bin/nutch solrindex http://localhost:8983/solr crawl/crawldb/
>>>>>
>>>> -linkdb
>>>
>>>> crawl/linkdb/ crawl/segments/20150406231502/ -filter -normalize"
>>>>>
>>>>> I always get a error
>>>>>
>>>>> Indexer: java.io.IOException: Job failed!
>>>>> at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1357)
>>>>> at org.apache.nutch.indexer.IndexingJob.index(IndexingJob.java:114)
>>>>> at org.apache.nutch.indexer.IndexingJob.run(IndexingJob.java:176)
>>>>> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
>>>>> at org.apache.nutch.indexer.IndexingJob.main(IndexingJob.java:186)
>>>>>
>>>>>
>>>>>
>>>>> Here is the complete hadoop log for the process.I have underlined the
>>>>>
>>>> error
>>>>
>>>>> part in it.
>>>>>
>>>>> 2015-04-07 09:38:06,613 INFO  indexer.IndexingJob - Indexer: starting
>>>>>
>>>> at
>>>
>>>> 2015-04-07 09:38:06
>>>>> 2015-04-07 09:38:06,684 INFO  indexer.IndexingJob - Indexer: deleting
>>>>>
>>>> gone
>>>>
>>>>> documents: false
>>>>> 2015-04-07 09:38:06,685 INFO  indexer.IndexingJob - Indexer: URL
>>>>>
>>>> filtering:
>>>>
>>>>> true
>>>>> 2015-04-07 09:38:06,685 INFO  indexer.IndexingJob - Indexer: URL
>>>>> normalizing: true
>>>>> 2015-04-07 09:38:06,893 INFO  indexer.IndexWriters - Adding
>>>>> org.apache.nutch.indexwriter.solr.SolrIndexWriter
>>>>> 2015-04-07 09:38:06,893 INFO  indexer.IndexingJob - Active
>>>>>
>>>> IndexWriters :
>>>
>>>> SOLRIndexWriter
>>>>> solr.server.url : URL of the SOLR instance (mandatory)
>>>>> solr.commit.size : buffer size when sending to SOLR (default 1000)
>>>>> solr.mapping.file : name of the mapping file for fields (default
>>>>> solrindex-mapping.xml)
>>>>> solr.auth : use authentication (default false)
>>>>> solr.auth.username : use authentication (default false)
>>>>> solr.auth : username for authentication
>>>>> solr.auth.password : password for authentication
>>>>>
>>>>>
>>>>> 2015-04-07 09:38:06,898 INFO  indexer.IndexerMapReduce -
>>>>>
>>>> IndexerMapReduce:
>>>>
>>>>> crawldb: crawl/crawldb
>>>>> 2015-04-07 09:38:06,898 INFO  indexer.IndexerMapReduce -
>>>>>
>>>> IndexerMapReduce:
>>>>
>>>>> linkdb: crawl/linkdb
>>>>> 2015-04-07 09:38:06,898 INFO  indexer.IndexerMapReduce -
>>>>>
>>>> IndexerMapReduces:
>>>>
>>>>> adding segment: crawl/segments/20150406231502
>>>>> 2015-04-07 09:38:07,036 WARN  util.NativeCodeLoader - Unable to load
>>>>> native-hadoop library for your platform... using builtin-java classes
>>>>>
>>>> where
>>>>
>>>>> applicable
>>>>> 2015-04-07 09:38:07,540 INFO  anchor.AnchorIndexingFilter - Anchor
>>>>> deduplication is: off
>>>>> 2015-04-07 09:38:07,565 INFO  regex.RegexURLNormalizer - can't find
>>>>>
>>>> rules
>>>
>>>> for scope 'indexer', using default
>>>>> 2015-04-07 09:38:09,552 INFO  regex.RegexURLNormalizer - can't find
>>>>>
>>>> rules
>>>
>>>> for scope 'indexer', using default
>>>>> 2015-04-07 09:38:10,642 INFO  regex.RegexURLNormalizer - can't find
>>>>>
>>>> rules
>>>
>>>> for scope 'indexer', using default
>>>>> 2015-04-07 09:38:10,734 INFO  regex.RegexURLNormalizer - can't find
>>>>>
>>>> rules
>>>
>>>> for scope 'indexer', using default
>>>>> 2015-04-07 09:38:10,895 INFO  regex.RegexURLNormalizer - can't find
>>>>>
>>>> rules
>>>
>>>> for scope 'indexer', using default
>>>>> 2015-04-07 09:38:11,088 INFO  regex.RegexURLNormalizer - can't find
>>>>>
>>>> rules
>>>
>>>> for scope 'indexer', using default
>>>>> 2015-04-07 09:38:11,219 INFO  indexer.IndexWriters - Adding
>>>>> org.apache.nutch.indexwriter.solr.SolrIndexWriter
>>>>> 2015-04-07 09:38:11,237 INFO  solr.SolrMappingReader - source: content
>>>>> dest: content
>>>>> 2015-04-07 09:38:11,237 INFO  solr.SolrMappingReader - source: title
>>>>>
>>>> dest:
>>>>
>>>>> title
>>>>> 2015-04-07 09:38:11,237 INFO  solr.SolrMappingReader - source: host
>>>>>
>>>> dest:
>>>
>>>> host
>>>>> 2015-04-07 09:38:11,237 INFO  solr.SolrMappingReader - source: segment
>>>>> dest: segment
>>>>> 2015-04-07 09:38:11,237 INFO  solr.SolrMappingReader - source: boost
>>>>>
>>>> dest:
>>>>
>>>>> boost
>>>>> 2015-04-07 09:38:11,237 INFO  solr.SolrMappingReader - source: digest
>>>>>
>>>> dest:
>>>>
>>>>> digest
>>>>> 2015-04-07 09:38:11,237 INFO  solr.SolrMappingReader - source: tstamp
>>>>>
>>>> dest:
>>>>
>>>>> tstamp
>>>>> 2015-04-07 09:38:11,526 INFO  solr.SolrIndexWriter - Indexing 250
>>>>>
>>>> documents
>>>>
>>>>> 2015-04-07 09:38:11,526 INFO  solr.SolrIndexWriter - Deleting 0
>>>>>
>>>> documents
>>>
>>>> 2015-04-07 09:38:11,644 INFO  solr.SolrIndexWriter - Indexing 250
>>>>>
>>>> documents
>>>>
>>>>> *2015-04-07 09:38:11,699 WARN  mapred.LocalJobRunner -
>>>>> job_local1245074757_0001*
>>>>> *org.apache.solr.common.SolrException: Not Found*
>>>>>
>>>>> *Not Found*
>>>>>
>>>>> *request: http://localhost:8983/solr/update?wt=javabin&version=2
>>>>> <http://localhost:8983/solr/update?wt=javabin&version=2>*
>>>>> * at
>>>>> org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.
>>>>>
>>>> request(CommonsHttpSolrServer.java:430)*
>>>>
>>>>> * at
>>>>> org.apache.solr.client.solrj.impl.CommonsHttpSolrServer.
>>>>>
>>>> request(CommonsHttpSolrServer.java:244)*
>>>>
>>>>> * at
>>>>> org.apache.solr.client.solrj.request.AbstractUpdateRequest.
>>>>>
>>>> process(AbstractUpdateRequest.java:105)*
>>>>
>>>>> * at
>>>>> org.apache.nutch.indexwriter.solr.SolrIndexWriter.write(
>>>>>
>>>> SolrIndexWriter.java:135)*
>>>>
>>>>> * at org.apache.nutch.indexer.IndexWriters.write(
>>>>> IndexWriters.java:88)*
>>>>> * at
>>>>> org.apache.nutch.indexer.IndexerOutputFormat$1.write(
>>>>>
>>>> IndexerOutputFormat.java:50)*
>>>>
>>>>> * at
>>>>> org.apache.nutch.indexer.IndexerOutputFormat$1.write(
>>>>>
>>>> IndexerOutputFormat.java:41)*
>>>>
>>>>> * at
>>>>> org.apache.hadoop.mapred.ReduceTask$OldTrackingRecordWriter.write(
>>>>>
>>>> ReduceTask.java:458)*
>>>>
>>>>> * at
>>>>>
>>>> org.apache.hadoop.mapred.ReduceTask$3.collect(ReduceTask.java:500)*
>>>
>>>> * at
>>>>> org.apache.nutch.indexer.IndexerMapReduce.reduce(
>>>>>
>>>> IndexerMapReduce.java:323)*
>>>>
>>>>> * at
>>>>> org.apache.nutch.indexer.IndexerMapReduce.reduce(
>>>>>
>>>> IndexerMapReduce.java:53)*
>>>>
>>>>> * at org.apache.hadoop.mapred.ReduceTask.runOldReducer(
>>>>>
>>>> ReduceTask.java:522)*
>>>>
>>>>> * at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:421)*
>>>>> * at
>>>>> org.apache.hadoop.mapred.LocalJobRunner$Job.run(
>>>>>
>>>> LocalJobRunner.java:398)*
>>>>
>>>>> *2015-04-07 09:38:12,408 ERROR indexer.IndexingJob - Indexer:
>>>>> java.io.IOException: Job failed!*
>>>>> * at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1357)*
>>>>> * at org.apache.nutch.indexer.IndexingJob.index(IndexingJob.java:114)*
>>>>> * at org.apache.nutch.indexer.IndexingJob.run(IndexingJob.java:176)*
>>>>> * at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)*
>>>>> * at org.apache.nutch.indexer.IndexingJob.main(IndexingJob.java:186)*
>>>>>
>>>>>> On Tue, 7 Apr 2015 at 03:18 Jeff Cocking <[email protected]>
>>>>>>
>>>>> wrote:
>>>>
>>>>> With Solr5.0.0 you can skip that step.  Solr will auto create your
>>>>>>
>>>>> schema
>>>>
>>>>> document based on the data being provided.
>>>>>>
>>>>>> One of the new features with Solr5 is the install/service feature. I
>>>>>>
>>>>> did a
>>>>
>>>>> quick write up on how to install Solr5 on Centos.  Might be something
>>>>>> useful there for you.
>>>>>>
>>>>>> http://www.cocking.com/apache-solr-5-0-install-on-centos-7/
>>>>>>
>>>>>> jeff
>>>>>>
>>>>>> On Mon, Apr 6, 2015 at 3:13 PM, Anchit Jain <[email protected]
>>>>>> wrote:
>>>>>>
>>>>>>  I want to index nutch results using *Solr 5.0* but as mentioned in
>>>>>>> https://wiki.apache.org/nutch/NutchTutorial there is no directory
>>>>>>> ${APACHE_SOLR_HOME}/example/solr/collection1/conf/
>>>>>>> in  solr 5.0 . So where I have to copy *schema.xml*?
>>>>>>> Also there is no *start.jar* present in example directory.
>>>>>>>
>>>>>>
>

Reply via email to