A field _indexed_at_tdt added when I index documents.

2015-12-17 Thread Guillermo Ortiz
I'm indexing documents in solr with Spark and it's missing the a field
 _indexed_at_tdt who is doesn't exist in my documents.

I have added this field in my schema, why is this field being added? any
solution?


Trying to index document in Solr with solr-spark library

2015-12-16 Thread Guillermo Ortiz
I'm getting some errors when I try to use the solr-sparl library getting
the error *KeeperErrorCode = NoNode for /live_nodes*.

I download the library and compile with the branch_4.x since I'm using
Cloudera 5.5.1 and Solr 4.10.3.

I checked the logs of Solr and Zookeeper and I didn't find any error and
navigate inside Zookeeper and the collection is created. These errors
happen in the executors of Spark.


2015-12-16 16:31:43,923 [Executor task launch worker-1] INFO
org.apache.zookeeper.ZooKeeper - Session: 0x1519126c7d55b23 closed

2015-12-16 16:31:43,924 [Executor task launch worker-1] ERROR org.apache.
spark.executor.Executor - Exception in task 5.2 in stage 12.0 (TID 218)
org.apache.solr.common.cloud.ZooKeeperException:
at org.apache.solr
.client.solrj.impl.CloudSolrServer.connect(CloudSolrServer.java:252)
at com.lucidworks.spark.SolrSupport.getSolrServer(SolrSupport.java:67)
at com.lucidworks.spark.SolrSupport$4.call(SolrSupport.java:162)
at com.lucidworks.spark.SolrSupport$4.call(SolrSupport.java:160)
at org.apache.spark
.api.java.JavaRDDLike$$anonfun$foreachPartition$1.apply(JavaRDDLike.scala:222)
at org.apache.spark
.api.java.JavaRDDLike$$anonfun$foreachPartition$1.apply(JavaRDDLike.scala:222)
at org.apache.spark
.rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$29.apply(RDD.scala:898)
at org.apache.spark
.rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$29.apply(RDD.scala:898)
at org.apache.spark
.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1850)
at org.apache.spark
.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:1850)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:88)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
*Caused by: org.apache.zookeeper.KeeperException$NoNodeException:
KeeperErrorCode = NoNode for /live_nodes*
at org.apache.zookeeper.KeeperException.create(KeeperException.java:111)
at org.apache.zookeeper.KeeperException.create(KeeperException.java:51)
at org.apache.zookeeper.ZooKeeper.getChildren(ZooKeeper.java:1468)
at org.apache.solr
.common.cloud.SolrZkClient$7.execute(SolrZkClient.java:290)
at org.apache.solr
.common.cloud.SolrZkClient$7.execute(SolrZkClient.java:287)
at org.apache.solr
.common.cloud.ZkCmdExecutor.retryOperation(ZkCmdExecutor.java:74)
at org.apache.solr
.common.cloud.SolrZkClient.getChildren(SolrZkClient.java:287)
at org.apache.solr
.common.cloud.ZkStateReader.createClusterStateWatchersAndUpdate(ZkStateReader.java:334)
at org.apache.solr
.client.solrj.impl.CloudSolrServer.connect(CloudSolrServer.java:243)