Looks like your Spark job is not connecting to the same Zookeeper as your Solr nodes.
Or, I suppose, the Solr nodes aren't started. You might get more information on the Cloudera help boards Best, Erick On Wed, Dec 16, 2015 at 11:58 PM, Guillermo Ortiz <konstt2...@gmail.com> wrote: > I'm getting some errors when I try to use the solr-sparl library getting > the error *KeeperErrorCode = NoNode for /live_nodes*. > > I download the library and compile with the branch_4.x since I'm using > Cloudera 5.5.1 and Solr 4.10.3. > > I checked the logs of Solr and Zookeeper and I didn't find any error and > navigate inside Zookeeper and the collection is created. These errors > happen in the executors of Spark. > > > 2015-12-16 16:31:43,923 [Executor task launch worker-1] INFO > org.apache.zookeeper.ZooKeeper - Session: 0x1519126c7d55b23 closed > > 2015-12-16 16:31:43,924 [Executor task launch worker-1] ERROR org.apache. > spark.executor.Executor - Exception in task 5.2 in stage 12.0 (TID 218) > org.apache.solr.common.cloud.ZooKeeperException: > at org.apache.solr > .client.solrj.impl.CloudSolrServer.connect(CloudSolrServer.java:252) > at com.lucidworks.spark.SolrSupport.getSolrServer(SolrSupport.java:67) > at com.lucidworks.spark.SolrSupport$4.call(SolrSupport.java:162) > at com.lucidworks.spark.SolrSupport$4.call(SolrSupport.java:160) > at org.apache.spark > .api.java.JavaRDDLike$$anonfun$foreachPartition$1.apply(JavaRDDLike.scala:222) > at org.apache.spark > .api.java.JavaRDDLike$$anonfun$foreachPartition$1.apply(JavaRDDLike.scala:222) > at org.apache.spark > .rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$29.apply(RDD.scala:898) > at org.apache.spark > .rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$29.apply(RDD.scala:898) > at org.apache.spark > .SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1850) > at org.apache.spark > .SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1850) > at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) > at org.apache.spark.scheduler.Task.run(Task.scala:88) > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:745) > *Caused by: org.apache.zookeeper.KeeperException$NoNodeException: > KeeperErrorCode = NoNode for /live_nodes* > at org.apache.zookeeper.KeeperException.create(KeeperException.java:111) > at org.apache.zookeeper.KeeperException.create(KeeperException.java:51) > at org.apache.zookeeper.ZooKeeper.getChildren(ZooKeeper.java:1468) > at org.apache.solr > .common.cloud.SolrZkClient$7.execute(SolrZkClient.java:290) > at org.apache.solr > .common.cloud.SolrZkClient$7.execute(SolrZkClient.java:287) > at org.apache.solr > .common.cloud.ZkCmdExecutor.retryOperation(ZkCmdExecutor.java:74) > at org.apache.solr > .common.cloud.SolrZkClient.getChildren(SolrZkClient.java:287) > at org.apache.solr > .common.cloud.ZkStateReader.createClusterStateWatchersAndUpdate(ZkStateReader.java:334) > at org.apache.solr > .client.solrj.impl.CloudSolrServer.connect(CloudSolrServer.java:243)