Hello, I am attempting to integrate Nutch 1.13 with Solr 6.6 running in cloud mode. I previously had this working fine with Nutch 1.13 and Solr 6.5 running in stand-alone mode, but now I get an error. It seems to be an issue with the collection not being default.
Command: bin/nutch index -Dsolr.zookeeper.hosts=localhost:9983 -Dsolr.auth.password=xxxxxxxx -Dsolr.auth.username=xxxxxxxx -Dsolr.auth=true -Dsolr.server.url=http://localhost:8983/solr/uc_website crawl/crawldb -linkdb crawl/linkdb crawl/segments/20170607135140 Result in hadoop.log: java.lang.Exception: java.io.IOException at org.apache.hadoop.mapred.LocalJobRunner$Job.runTasks(LocalJobRunner.java:462) at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:529) Caused by: java.io.IOException at org.apache.nutch.indexwriter.solr.SolrIndexWriter.makeIOException(SolrIndexWriter.java:234) at org.apache.nutch.indexwriter.solr.SolrIndexWriter.push(SolrIndexWriter.java:213) at org.apache.nutch.indexwriter.solr.SolrIndexWriter.write(SolrIndexWriter.java:174) at org.apache.nutch.indexer.IndexWriters.write(IndexWriters.java:87) at org.apache.nutch.indexer.IndexerOutputFormat$1.write(IndexerOutputFormat.java:50) at org.apache.nutch.indexer.IndexerOutputFormat$1.write(IndexerOutputFormat.java:41) at org.apache.hadoop.mapred.ReduceTask$OldTrackingRecordWriter.write(ReduceTask.java:493) at org.apache.hadoop.mapred.ReduceTask$3.collect(ReduceTask.java:422) at org.apache.nutch.indexer.IndexerMapReduce.reduce(IndexerMapReduce.java:368) at org.apache.nutch.indexer.IndexerMapReduce.reduce(IndexerMapReduce.java:57) at org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:444) at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:392) at org.apache.hadoop.mapred.LocalJobRunner$Job$ReduceTaskRunnable.run(LocalJobRunner.java:319) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) Caused by: org.apache.solr.client.solrj.SolrServerException: No collection param specified on request and no default collection has been set. at org.apache.solr.client.solrj.impl.CloudSolrClient.directUpdate(CloudSolrClient.java:556) at org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:981) at org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:870) at org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:806) at org.apache.solr.client.solrj.SolrClient.request(SolrClient.java:1219) at org.apache.nutch.indexwriter.solr.SolrIndexWriter.push(SolrIndexWriter.java:210) ... 16 more 2017-06-07 14:42:32,305 ERROR indexer.IndexingJob - Indexer: java.io.IOException: Job failed! at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:865) at org.apache.nutch.indexer.IndexingJob.index(IndexingJob.java:147) at org.apache.nutch.indexer.IndexingJob.run(IndexingJob.java:230) at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) at org.apache.nutch.indexer.IndexingJob.main(IndexingJob.java:239) I think the root of the problem is the line "No collection param specified on request and no default collection has been set." Any help is greatly appreciated. Thanks! -- Dave Parker Database & Systems Administrator Utica College Integrated Information Technology Services (315) 792-3229 Registered Linux User #408177

