Looks like your Spark job is not connecting to the same Zookeeper
as your Solr nodes.

Or, I suppose, the Solr nodes aren't started.

You might get more information on the Cloudera help boards

Best,
Erick

On Wed, Dec 16, 2015 at 11:58 PM, Guillermo Ortiz <konstt2...@gmail.com> wrote:
> I'm getting some errors when I try to use the solr-sparl library getting
> the error *KeeperErrorCode = NoNode for /live_nodes*.
>
> I download the library and compile with the branch_4.x since I'm using
> Cloudera 5.5.1 and Solr 4.10.3.
>
> I checked the logs of Solr and Zookeeper and I didn't find any error and
> navigate inside Zookeeper and the collection is created. These errors
> happen in the executors of Spark.
>
>
> 2015-12-16 16:31:43,923 [Executor task launch worker-1] INFO
> org.apache.zookeeper.ZooKeeper - Session: 0x1519126c7d55b23 closed
>
> 2015-12-16 16:31:43,924 [Executor task launch worker-1] ERROR org.apache.
> spark.executor.Executor - Exception in task 5.2 in stage 12.0 (TID 218)
> org.apache.solr.common.cloud.ZooKeeperException:
> at org.apache.solr
> .client.solrj.impl.CloudSolrServer.connect(CloudSolrServer.java:252)
> at com.lucidworks.spark.SolrSupport.getSolrServer(SolrSupport.java:67)
> at com.lucidworks.spark.SolrSupport$4.call(SolrSupport.java:162)
> at com.lucidworks.spark.SolrSupport$4.call(SolrSupport.java:160)
> at org.apache.spark
> .api.java.JavaRDDLike$$anonfun$foreachPartition$1.apply(JavaRDDLike.scala:222)
> at org.apache.spark
> .api.java.JavaRDDLike$$anonfun$foreachPartition$1.apply(JavaRDDLike.scala:222)
> at org.apache.spark
> .rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$29.apply(RDD.scala:898)
> at org.apache.spark
> .rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$29.apply(RDD.scala:898)
> at org.apache.spark
> .SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1850)
> at org.apache.spark
> .SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1850)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
> at org.apache.spark.scheduler.Task.run(Task.scala:88)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> *Caused by: org.apache.zookeeper.KeeperException$NoNodeException:
> KeeperErrorCode = NoNode for /live_nodes*
> at org.apache.zookeeper.KeeperException.create(KeeperException.java:111)
> at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
> at org.apache.zookeeper.ZooKeeper.getChildren(ZooKeeper.java:1468)
> at org.apache.solr
> .common.cloud.SolrZkClient$7.execute(SolrZkClient.java:290)
> at org.apache.solr
> .common.cloud.SolrZkClient$7.execute(SolrZkClient.java:287)
> at org.apache.solr
> .common.cloud.ZkCmdExecutor.retryOperation(ZkCmdExecutor.java:74)
> at org.apache.solr
> .common.cloud.SolrZkClient.getChildren(SolrZkClient.java:287)
> at org.apache.solr
> .common.cloud.ZkStateReader.createClusterStateWatchersAndUpdate(ZkStateReader.java:334)
> at org.apache.solr
> .client.solrj.impl.CloudSolrServer.connect(CloudSolrServer.java:243)

Reply via email to