Hi, Everyone if you check the log file it does talk about the error, here is re-briefing of problem.
1. Solr without any authentication => Nutch work successfully and it populate the solr core say (abc) 2. Solr with protection and Nutch solr.auth=false => unauthorized access which make sense. 3. Solr with protection and Nutch solr.auth=trur and correct id and pass in config file => It spit out the error and I have attached the log at the bottom of this email. When I use authentication nutch is not able to insert data. However problem is not related to solr because if I try to populate data with solr having id and password and nutch with solr.auth=false it does print unauthorized access and that makes sense. Now with solr.auth=true and id and password in nutch-default nutch is not able to insert data and below is the error log. I guess is there any user right like admin or content-admin in solr ?? That too I tried with all kind of users and always same error. If some one can try and see if they can push the data with protected solr. If you are not getting error then please tell me what are the configuration you are using in detail ?? Treat me like novice and then tell me how to do it. Because I tried all kind of permutation of configuration both in solr and nutch side without any luck. Please do help me this is a genuine request . I do understand you guys are pretty busy with your work its not that i am just bothering you without my homework. Please see the log 2016-01-20 07:02:15,658 INFO indexer.IndexWriters - Adding org.apache.nutch.indexwriter.solr.SolrIndexWriter 2016-01-20 07:04:36,366 WARN util.NativeCodeLoader - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2016-01-20 07:04:36,656 INFO segment.SegmentChecker - Segment dir is complete: file:/home/rakesh/Desktop/arima/nutch/runtime/local/yahCrawl/segements/20160119163402. 2016-01-20 07:04:36,658 INFO segment.SegmentChecker - Segment dir is complete: file:/home/rakesh/Desktop/arima/nutch/runtime/local/yahCrawl/segements/20160119163656. 2016-01-20 07:04:36,673 INFO segment.SegmentChecker - Segment dir is complete: file:/home/rakesh/Desktop/arima/nutch/runtime/local/yahCrawl/segements/20160119164952. 2016-01-20 07:04:36,674 INFO indexer.IndexingJob - Indexer: starting at 2016-01-20 07:04:36 2016-01-20 07:04:36,676 INFO indexer.IndexingJob - Indexer: deleting gone documents: false 2016-01-20 07:04:36,676 INFO indexer.IndexingJob - Indexer: URL filtering: false 2016-01-20 07:04:36,676 INFO indexer.IndexingJob - Indexer: URL normalizing: false 2016-01-20 07:04:37,036 INFO indexer.IndexWriters - Adding org.apache.nutch.indexwriter.solr.SolrIndexWriter 2016-01-20 07:04:37,036 INFO indexer.IndexingJob - Active IndexWriters : SolrIndexWriter solr.server.type : Type of SolrServer to communicate with (default 'http' however options include 'cloud', 'lb' and 'concurrent') solr.server.url : URL of the Solr instance (mandatory) solr.zookeeper.url : URL of the Zookeeper URL (mandatory if 'cloud' value for solr.server.type) solr.loadbalance.urls : Comma-separated string of Solr server strings to be used (madatory if 'lb' value for solr.server.type) solr.mapping.file : name of the mapping file for fields (default solrindex-mapping.xml) solr.commit.size : buffer size when sending to Solr (default 1000) solr.auth : use authentication (default false) solr.auth.username : username for authentication solr.auth.password : password for authentication 2016-01-20 07:04:37,039 INFO indexer.IndexerMapReduce - IndexerMapReduce: crawldb: yahCrawl/crawldb 2016-01-20 07:04:37,039 INFO indexer.IndexerMapReduce - IndexerMapReduce: linkdb: yahCrawl/linkdb 2016-01-20 07:04:37,039 INFO indexer.IndexerMapReduce - IndexerMapReduces: adding segment: file:/home/rakesh/Desktop/arima/nutch/runtime/local/yahCrawl/segements/20160119163402 2016-01-20 07:04:37,045 INFO indexer.IndexerMapReduce - IndexerMapReduces: adding segment: file:/home/rakesh/Desktop/arima/nutch/runtime/local/yahCrawl/segements/20160119163656 2016-01-20 07:04:37,046 INFO indexer.IndexerMapReduce - IndexerMapReduces: adding segment: file:/home/rakesh/Desktop/arima/nutch/runtime/local/yahCrawl/segements/20160119164952 2016-01-20 07:04:37,047 WARN indexer.IndexerMapReduce - Ignoring linkDb for indexing, no linkDb found in path: yahCrawl/linkdb 2016-01-20 07:04:38,151 WARN conf.Configuration - file:/tmp/hadoop-rakesh/mapred/staging/rakesh1643615475/.staging/job_local1643615475_0001/job.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.retry.interval; Ignoring. 2016-01-20 07:04:38,153 WARN conf.Configuration - file:/tmp/hadoop-rakesh/mapred/staging/rakesh1643615475/.staging/job_local1643615475_0001/job.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.attempts; Ignoring. 2016-01-20 07:04:38,312 WARN conf.Configuration - file:/tmp/hadoop-rakesh/mapred/local/localRunner/rakesh/job_local1643615475_0001/job_local1643615475_0001.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.retry.interval; Ignoring. 2016-01-20 07:04:38,314 WARN conf.Configuration - file:/tmp/hadoop-rakesh/mapred/local/localRunner/rakesh/job_local1643615475_0001/job_local1643615475_0001.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.attempts; Ignoring. 2016-01-20 07:04:39,258 INFO anchor.AnchorIndexingFilter - Anchor deduplication is: off 2016-01-20 07:04:40,773 INFO indexer.IndexWriters - Adding org.apache.nutch.indexwriter.solr.SolrIndexWriter *2016-01-20 07:04:40,784 INFO solr.SolrUtils - Authenticating as: radmin* 2016-01-20 07:04:41,018 INFO solr.SolrMappingReader - source: content dest: content 2016-01-20 07:04:41,018 INFO solr.SolrMappingReader - source: title dest: title 2016-01-20 07:04:41,018 INFO solr.SolrMappingReader - source: host dest: host 2016-01-20 07:04:41,018 INFO solr.SolrMappingReader - source: segment dest: segment 2016-01-20 07:04:41,018 INFO solr.SolrMappingReader - source: boost dest: boost 2016-01-20 07:04:41,018 INFO solr.SolrMappingReader - source: digest dest: digest 2016-01-20 07:04:41,018 INFO solr.SolrMappingReader - source: tstamp dest: tstamp 2016-01-20 07:04:41,091 INFO solr.SolrIndexWriter - Indexing 3 documents 2016-01-20 07:04:41,340 INFO solr.SolrIndexWriter - Indexing 3 documents 2016-01-20 07:04:41,398 WARN mapred.LocalJobRunner - job_local1643615475_0001 java.lang.Exception: java.io.IOException at org.apache.hadoop.mapred.LocalJobRunner$Job.runTasks(LocalJobRunner.java:462) at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:529) Caused by: java.io.IOException at org.apache.nutch.indexwriter.solr.SolrIndexWriter.makeIOException(SolrIndexWriter.java:171) at org.apache.nutch.indexwriter.solr.SolrIndexWriter.close(SolrIndexWriter.java:157) at org.apache.nutch.indexer.IndexWriters.close(IndexWriters.java:115) at org.apache.nutch.indexer.IndexerOutputFormat$1.close(IndexerOutputFormat.java:44) at org.apache.hadoop.mapred.ReduceTask$OldTrackingRecordWriter.close(ReduceTask.java:502) at org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:456) at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:392) at org.apache.hadoop.mapred.LocalJobRunner$Job$ReduceTaskRunnable.run(LocalJobRunner.java:319) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) Caused by: org.apache.solr.client.solrj.SolrServerException: IOException occured when talking to server at: http://127.0.0.1:8983/solr/mah at org.apache.solr.client.solrj.impl.HttpSolrServer.executeMethod(HttpSolrServer.java:566) at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:210) at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:206) at org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:124) at org.apache.nutch.indexwriter.solr.SolrIndexWriter.close(SolrIndexWriter.java:153) ... 11 more *Caused by: org.apache.http.client.ClientProtocolException* at org.apache.http.impl.client.InternalHttpClient.doExecute(InternalHttpClient.java:186) at org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:82) at org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:106) at org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:57) at org.apache.solr.client.solrj.impl.HttpSolrServer.executeMethod(HttpSolrServer.java:448) ... 15 more Caused by: org.apache.http.client.NonRepeatableRequestException: Cannot retry request with a non-repeatable request entity. at org.apache.http.impl.execchain.MainClientExec.execute(MainClientExec.java:208) at org.apache.http.impl.execchain.ProtocolExec.execute(ProtocolExec.java:195) at org.apache.http.impl.execchain.RetryExec.execute(RetryExec.java:86) at org.apache.http.impl.execchain.RedirectExec.execute(RedirectExec.java:108) at org.apache.http.impl.client.InternalHttpClient.doExecute(InternalHttpClient.java:184) ... 19 more 2016-01-20 07:04:42,430 ERROR indexer.IndexingJob - Indexer: java.io.IOException: Job failed! at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:836) at org.apache.nutch.indexer.IndexingJob.index(IndexingJob.java:145) at org.apache.nutch.indexer.IndexingJob.run(IndexingJob.java:228) at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) at org.apache.nutch.indexer.IndexingJob.main(IndexingJob.java:237) On Tue, Jan 19, 2016 at 7:44 PM, Roannel Fernández Hernández <[email protected] > wrote: > Hi > > I think that your problem is not related with Solr authentication. The > fields of documents sent by you to Solr and the fields defined in Solr > schema are differents. Perhaps the Nutch document has a multivalued field > defined in Solr schema as simple field, or in Solr schema there is a > required field not sent by Nutch or the primary key has not been sent or ... > > Just to confirm that and if it's possible you can remove the Solr > protection and try it again. If you get the same error, then it is not > related with Solr authentication and you have to check the fields sent to > Solr. > > Regards > > > ------------------------------ > > *From: *"Zara Parst" <[email protected]> > *To: *[email protected] > *Sent: *Monday, January 18, 2016 3:28:29 PM > *Subject: *[MASSMAIL]Re: Nutch/Solr communication problem > > > I am using solr 5.4 and nutch 1.11 > > On Tue, Jan 19, 2016 at 1:46 AM, Markus Jelsma <[email protected] > > wrote: > >> Hi - it was an answer to your question whether i have ever used it. Yes, >> i patched and committed it. And therefore i asked if you're using Solr 5 or >> not. So again, are you using Solr 5? >> >> Markus >> >> >> -----Original message----- >> From: Zara Parst<[email protected]> >> Sent: Monday 18th January 2016 16:16 >> To: [email protected] >> Subject: Re: Nutch/Solr communication problem >> >> Mind to share that patch ? >> >> On Mon, Jan 18, 2016 at 8:28 PM, Markus Jelsma < >> [email protected] <mailto:[email protected]>> wrote: >> Yes i have used it, i made the damn patch myself years ago, and i used >> the same configuration. Command line or config work the same. >> >> Markus >> >> -----Original message----- >> >> From: Zara Parst<[email protected] <mailto:[email protected]>> >> >> Sent: Monday 18th January 2016 12:55 >> >> To: [email protected] <mailto:[email protected]> >> >> Subject: Re: Nutch/Solr communication problem >> >> Dear Markus, >> >> Are you just speaking blindly or what ?? My concern is did you ever try >> pushing index to solr which is password protected ? If yes can you just >> tell me what were the config you used , if you did that in config file then >> let me know or if you did through command then please let me know. >> >> thanks >> >> On Mon, Jan 18, 2016 at 4:50 PM, Markus Jelsma < >> [email protected] <mailto:[email protected]> <mailto: >> [email protected] <mailto:[email protected]>>> wrote: >> >> Hi - This doesnt look like a HTTP basic authentication problem. Are you >> running Solr 5.x? >> >> Markus >> >> -----Original message----- >> >> From: Zara Parst<[email protected] <mailto:[email protected]> >> <mailto:[email protected] <mailto:[email protected]>>> >> >> Sent: Monday 18th January 2016 11:55 >> >> To: [email protected] <mailto:[email protected]> <mailto: >> [email protected] <mailto:[email protected]>> >> >> Subject: Re: Nutch/Solr communication problem >> >> SolrIndexWriter >> >> solr.server.type : Type of SolrServer to communicate with >> (default http however options include cloud, lb and concurrent) >> >> solr.server.url : URL of the Solr instance (mandatory) >> >> solr.zookeeper.url : URL of the Zookeeper URL (mandatory if cloud >> value for solr.server.type) >> >> solr.loadbalance.urls : Comma-separated string of Solr server >> strings to be used (madatory if lb value for solr.server.type) >> >> solr.mapping.file : name of the mapping file for fields (default >> solrindex-mapping.xml) >> >> solr.commit.size : buffer size when sending to Solr (default 1000) >> >> solr.auth : use authentication (default false) >> >> solr.auth.username : username for authentication >> >> solr.auth.password : password for authentication >> >> 2016-01-17 19:19:42,973 INFO indexer.IndexerMapReduce - >> IndexerMapReduce: crawldb: crawlDbyah/crawldb >> >> 2016-01-17 19:19:42,973 INFO indexer.IndexerMapReduce - >> IndexerMapReduce: linkdb: crawlDbyah/linkdb >> >> 2016-01-17 19:19:42,973 INFO indexer.IndexerMapReduce - >> IndexerMapReduces: adding segment: crawlDbyah/segments/20160117191906 >> >> 2016-01-17 19:19:42,975 WARN indexer.IndexerMapReduce - Ignoring linkDb >> for indexing, no linkDb found in path: crawlDbyah/linkdb >> >> 2016-01-17 19:19:43,807 WARN conf.Configuration - >> file:/tmp/hadoop-rakesh/mapred/staging/rakesh2114349538/.staging/job_local2114349538_0001/job.xml:an >> attempt to override final parameter: >> mapreduce.job.end-notification.max.retry.interval; Ignoring. >> >> 2016-01-17 19:19:43,809 WARN conf.Configuration - >> file:/tmp/hadoop-rakesh/mapred/staging/rakesh2114349538/.staging/job_local2114349538_0001/job.xml:an >> attempt to override final parameter: >> mapreduce.job.end-notification.max.attempts; Ignoring. >> >> 2016-01-17 19:19:43,963 WARN conf.Configuration - >> file:/tmp/hadoop-rakesh/mapred/local/localRunner/rakesh/job_local2114349538_0001/job_local2114349538_0001.xml:an >> attempt to override final parameter: >> mapreduce.job.end-notification.max.retry.interval; Ignoring. >> >> 2016-01-17 19:19:43,980 WARN conf.Configuration - >> file:/tmp/hadoop-rakesh/mapred/local/localRunner/rakesh/job_local2114349538_0001/job_local2114349538_0001.xml:an >> attempt to override final parameter: >> mapreduce.job.end-notification.max.attempts; Ignoring. >> >> 2016-01-17 19:19:44,260 INFO anchor.AnchorIndexingFilter - Anchor >> deduplication is: off >> >> 2016-01-17 19:19:45,128 INFO indexer.IndexWriters - Adding >> org.apache.nutch.indexwriter.solr.SolrIndexWriter >> >> 2016-01-17 19:19:45,148 INFO solr.SolrUtils - Authenticating as: radmin >> >> 2016-01-17 19:19:45,318 INFO solr.SolrMappingReader - source: content >> dest: content >> >> 2016-01-17 19:19:45,318 INFO solr.SolrMappingReader - source: title >> dest: title >> >> 2016-01-17 19:19:45,318 INFO solr.SolrMappingReader - source: host dest: >> host >> >> 2016-01-17 19:19:45,319 INFO solr.SolrMappingReader - source: segment >> dest: segment >> >> 2016-01-17 19:19:45,319 INFO solr.SolrMappingReader - source: boost >> dest: boost >> >> 2016-01-17 19:19:45,319 INFO solr.SolrMappingReader - source: digest >> dest: digest >> >> 2016-01-17 19:19:45,319 INFO solr.SolrMappingReader - source: tstamp >> dest: tstamp >> >> 2016-01-17 19:19:45,360 INFO solr.SolrIndexWriter - Indexing 2 documents >> >> 2016-01-17 19:19:45,507 INFO solr.SolrIndexWriter - Indexing 2 documents >> >> 2016-01-17 19:19:45,526 WARN mapred.LocalJobRunner - >> job_local2114349538_0001 >> >> java.lang.Exception: java.io.IOException >> >> at >> org.apache.hadoop.mapred.LocalJobRunner$Job.runTasks(LocalJobRunner.java:462) >> >> at >> org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:529) >> >> Caused by: java.io.IOException >> >> at >> org.apache.nutch.indexwriter.solr.SolrIndexWriter.makeIOException(SolrIndexWriter.java:171) >> >> at >> org.apache.nutch.indexwriter.solr.SolrIndexWriter.close(SolrIndexWriter.java:157) >> >> at >> org.apache.nutch.indexer.IndexWriters.close(IndexWriters.java:115) >> >> at >> org.apache.nutch.indexer.IndexerOutputFormat$1.close(IndexerOutputFormat.java:44) >> >> at >> org.apache.hadoop.mapred.ReduceTask$OldTrackingRecordWriter.close(ReduceTask.java:502) >> >> at >> org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:456) >> >> at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:392) >> >> at >> org.apache.hadoop.mapred.LocalJobRunner$Job$ReduceTaskRunnable.run(LocalJobRunner.java:319) >> >> at >> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) >> >> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >> >> at >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) >> >> at >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) >> >> at java.lang.Thread.run(Thread.java:745) >> >> Caused by: org.apache.solr.client.solrj.SolrServerException: IOException >> occured when talking to server at: http://127.0.0.1:8983/solr/yah < >> http://127.0.0.1:8983/solr/yah> <http://127.0.0.1:8983/solr/yah < >> http://127.0.0.1:8983/solr/yah>> <http://127.0.0.1:8983/solr/yah < >> http://127.0.0.1:8983/solr/yah> <http://127.0.0.1:8983/solr/yah < >> http://127.0.0.1:8983/solr/yah>>> >> >> at >> org.apache.solr.client.solrj.impl.HttpSolrServer.executeMethod(HttpSolrServer.java:566) >> >> at >> org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:210) >> >> at >> org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:206) >> >> at >> org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:124) >> >> at >> org.apache.nutch.indexwriter.solr.SolrIndexWriter.close(SolrIndexWriter.java:153) >> >> ... 11 more >> >> Caused by: org.apache.http.client.ClientProtocolException >> >> at >> org.apache.http.impl.client.InternalHttpClient.doExecute(InternalHttpClient.java:186) >> >> at >> org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:82) >> >> at >> org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:106) >> >> at >> org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:57) >> >> at >> org.apache.solr.client.solrj.impl.HttpSolrServer.executeMethod(HttpSolrServer.java:448) >> >> ... 15 more >> >> Caused by: org.apache.http.client.NonRepeatableRequestException: Cannot >> retry request with a non-repeatable request entity. >> >> at >> org.apache.http.impl.execchain.MainClientExec.execute(MainClientExec.java:208) >> >> at >> org.apache.http.impl.execchain.ProtocolExec.execute(ProtocolExec.java:195) >> >> at >> org.apache.http.impl.execchain.RetryExec.execute(RetryExec.java:86) >> >> at >> org.apache.http.impl.execchain.RedirectExec.execute(RedirectExec.java:108) >> >> at >> org.apache.http.impl.client.InternalHttpClient.doExecute(InternalHttpClient.java:184) >> >> ... 19 more >> >> 2016-01-17 19:19:46,055 ERROR indexer.IndexingJob - Indexer: >> java.io.IOException: Job failed! >> >> at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:836) >> >> at >> org.apache.nutch.indexer.IndexingJob.index(IndexingJob.java:145) >> >> at org.apache.nutch.indexer.IndexingJob.run(IndexingJob.java:228) >> >> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) >> >> at org.apache.nutch.indexer.IndexingJob.main(IndexingJob.java:237) >> >> On Mon, Jan 18, 2016 at 4:15 PM, Markus Jelsma < >> [email protected] <mailto:[email protected]> <mailto: >> [email protected] <mailto:[email protected]>> <mailto: >> [email protected] <mailto:[email protected]> <mailto: >> [email protected] <mailto:[email protected]>>>> wrote: >> >> Hi - can you post the log output? >> >> Markus >> >> -----Original message----- >> >> From: Zara Parst<[email protected] <mailto:[email protected]> >> <mailto:[email protected] <mailto:[email protected]>> <mailto: >> [email protected] <mailto:[email protected]> <mailto: >> [email protected] <mailto:[email protected]>>>> >> >> Sent: Monday 18th January 2016 2:06 >> >> To: [email protected] <mailto:[email protected]> <mailto: >> [email protected] <mailto:[email protected]>> <mailto: >> [email protected] <mailto:[email protected]> <mailto: >> [email protected] <mailto:[email protected]>>> >> >> Subject: Nutch/Solr communication problem >> >> Hi everyone, >> >> I have situation here, I am using nutch 1.11 and solr 5.4 >> >> Solr is protected by user name and password I am passing credential to >> solr using following command >> >> bin/crawl -i -Dsolr.server.url=http://localhost:8983/solr/abc < >> http://localhost:8983/solr/abc> <http://localhost:8983/solr/abc < >> http://localhost:8983/solr/abc>> <http://localhost:8983/solr/abc < >> http://localhost:8983/solr/abc> <http://localhost:8983/solr/abc < >> http://localhost:8983/solr/abc>>> <http://localhost:8983/solr/abc < >> http://localhost:8983/solr/abc> <http://localhost:8983/solr/abc < >> http://localhost:8983/solr/abc>> <http://localhost:8983/solr/abc < >> http://localhost:8983/solr/abc> <http://localhost:8983/solr/abc < >> http://localhost:8983/solr/abc>>>> -D solr.auth=true >> -Dsolr.auth.username=xxxx -Dsolr.auth.password=xxx url crawlDbyah 1 >> >> and always same problem , please help me how to feed data to protected >> solr. >> >> Below is error message. >> >> Indexer: starting at 2016-01-17 19:01:12 >> >> Indexer: deleting gone documents: false >> >> Indexer: URL filtering: false >> >> Indexer: URL normalizing: false >> >> Active IndexWriters : >> >> SolrIndexWriter >> >> solr.server.type : Type of SolrServer to communicate with >> (default http however options include cloud, lb and concurrent) >> >> solr.server.url : URL of the Solr instance (mandatory) >> >> solr.zookeeper.url : URL of the Zookeeper URL (mandatory if cloud >> value for solr.server.type) >> >> solr.loadbalance.urls : Comma-separated string of Solr server >> strings to be used (madatory if lb value for solr.server.type) >> >> solr.mapping.file : name of the mapping file for fields (default >> solrindex-mapping.xml) >> >> solr.commit.size : buffer size when sending to Solr (default 1000) >> >> solr.auth : use authentication (default false) >> >> solr.auth.username : username for authentication >> >> solr.auth.password : password for authentication >> >> Indexing 2 documents >> >> Indexing 2 documents >> >> Indexer: java.io.IOException: Job failed! >> >> at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:836) >> >> at >> org.apache.nutch.indexer.IndexingJob.index(IndexingJob.java:145) >> >> at org.apache.nutch.indexer.IndexingJob.run(IndexingJob.java:228) >> >> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) >> >> at org.apache.nutch.indexer.IndexingJob.main(IndexingJob.java:237) >> >> I also tried username and password in nutch-default.xml but again same >> error. Please help me out. >> >> >> > > > >

