and also make sure that hbase-site.xml is set in your classpath on all
nodes, both master and workers, and also client.

normally i put it into $SPARK_HOME/conf/ then the spark cluster will
be started with this conf file.

btw. @Ted, did you tried insert into hbase table with spark's
HiveContext? i got this issue:
https://issues.apache.org/jira/browse/SPARK-6628

and there is a patch available: https://issues.apache.org/jira/browse/HIVE-11166


2016-03-01 15:16 GMT+01:00 Ted Yu <yuzhih...@gmail.com>:
> 16/03/01 01:36:31 WARN TaskSetManager: Lost task 0.0 in stage 0.0 (TID 0,
> ip-xxx-xx-xx-xxx.ap-southeast-1.compute.internal):
> java.lang.RuntimeException: hbase-default.xml file seems to be for an older
> version of HBase (null), this version is 1.1.2.2.3.4.0-3485
>
> The above was likely caused by some component being built with different
> release of hbase.
>
> Try setting "hbase.defaults.for.version.skip" to true.
>
> Cheers
>
>
> On Mon, Feb 29, 2016 at 9:12 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>>
>> 16/02/29 23:09:34 INFO ZooKeeper: Initiating client connection,
>> connectString=localhost:2181 sessionTimeout=90000
>> watcher=hconnection-0x26fa89a20x0, quorum=localhost:2181, baseZNode=/hbase
>>
>> Since baseZNode didn't match what you set in hbase-site.xml, the cause was
>> likely that hbase-site.xml being inaccessible to your Spark job.
>>
>> Please add it in your classpath.
>>
>> On Mon, Feb 29, 2016 at 8:42 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>>>
>>> 16/02/29 23:09:34 INFO ClientCnxn: Opening socket connection to server
>>> localhost/0:0:0:0:0:0:0:1:2181. Will not attempt to authenticate using SASL
>>> (unknown error)
>>>
>>> Is your cluster secure cluster ?
>>>
>>> bq. Trace :
>>>
>>> Was there any output after 'Trace :' ?
>>>
>>> Was hbase-site.xml accessible to your Spark job ?
>>>
>>> Thanks
>>>
>>> On Mon, Feb 29, 2016 at 8:27 PM, Divya Gehlot <divya.htco...@gmail.com>
>>> wrote:
>>>>
>>>> Hi,
>>>> I am getting error when I am trying to connect hive table (which is
>>>> being created through HbaseIntegration) in spark
>>>>
>>>> Steps I followed :
>>>> Hive Table creation code  :
>>>> CREATE EXTERNAL TABLE IF NOT EXISTS TEST(NAME STRING,AGE INT)
>>>> STORED BY 'org.apache.hadoop.hive.hbase.HBaseStorageHandler'
>>>> WITH SERDEPROPERTIES ("hbase.columns.mapping" = ":key,0:AGE")
>>>> TBLPROPERTIES ("hbase.table.name" = "TEST",
>>>> "hbase.mapred.output.outputtable" = "TEST");
>>>>
>>>>
>>>> DESCRIBE TEST ;
>>>> col_name    data_type    comment
>>>> name            string         from deserializer
>>>> age               int             from deserializer
>>>>
>>>>
>>>> Spark Code :
>>>> import org.apache.spark._
>>>> import org.apache.spark.sql._
>>>>
>>>> val hiveContext = new org.apache.spark.sql.hive.HiveContext(sc)
>>>> hiveContext.sql("from TEST SELECT  NAME").collect.foreach(println)
>>>>
>>>>
>>>> Starting Spark shell
>>>> spark-shell --jars
>>>> /usr/hdp/2.3.4.0-3485/hive/lib/guava-14.0.1.jar,/usr/hdp/2.3.4.0-3485/hive/lib/hive-hbase-handler.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-client.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-common.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-protocol.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-server.jar,/usr/hdp/2.3.4.0-3485/hive/lib/hive-hbase-handler.jar,/usr/hdp/2.3.4.0-3485/hive/lib/htrace-core-3.1.0-incubating.jar,/usr/hdp/2.3.4.0-3485/hive/lib/zookeeper-3.4.6.2.3.4.0-3485.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-server.jar
>>>> --driver-class-path
>>>> /usr/hdp/2.3.4.0-3485/hive/lib/guava-14.0.1.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-client.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-common.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-protocol.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-server.jar,/usr/hdp/2.3.4.0-3485/hive/lib/hive-hbase-handler.jar,/usr/hdp/2.3.4.0-3485/hive/lib/htrace-core-3.1.0-incubating.jar,/usr/hdp/2.3.4.0-3485/hive/lib/zookeeper-3.4.6.2.3.4.0-3485.jar,/usr/hdp/2.3.4.0-3485/hive/lib/hive-hbase-handler.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-server.jar
>>>> --packages com.databricks:spark-csv_2.10:1.3.0  --master yarn-client -i
>>>> /TestDivya/Spark/InstrumentCopyToHDFSHive.scala
>>>>
>>>> Stack Trace :
>>>>
>>>>> Stack SQL context available as sqlContext.
>>>>> Loading /TestDivya/Spark/InstrumentCopyToHDFSHive.scala...
>>>>> import org.apache.spark._
>>>>> import org.apache.spark.sql._
>>>>> 16/02/29 23:09:29 INFO HiveContext: Initializing execution hive,
>>>>> version 1.2.1
>>>>> 16/02/29 23:09:29 INFO ClientWrapper: Inspected Hadoop version:
>>>>> 2.7.1.2.3.4.0-3485
>>>>> 16/02/29 23:09:29 INFO ClientWrapper: Loaded
>>>>> org.apache.hadoop.hive.shims.Hadoop23Shims for Hadoop version
>>>>> 2.7.1.2.3.4.0-3485
>>>>> 16/02/29 23:09:29 INFO HiveContext: default warehouse location is
>>>>> /user/hive/warehouse
>>>>> 16/02/29 23:09:29 INFO HiveContext: Initializing
>>>>> HiveMetastoreConnection version 1.2.1 using Spark classes.
>>>>> 16/02/29 23:09:29 INFO ClientWrapper: Inspected Hadoop version:
>>>>> 2.7.1.2.3.4.0-3485
>>>>> 16/02/29 23:09:29 INFO ClientWrapper: Loaded
>>>>> org.apache.hadoop.hive.shims.Hadoop23Shims for Hadoop version
>>>>> 2.7.1.2.3.4.0-3485
>>>>> 16/02/29 23:09:30 WARN NativeCodeLoader: Unable to load native-hadoop
>>>>> library for your platform... using builtin-java classes where applicable
>>>>> 16/02/29 23:09:30 INFO metastore: Trying to connect to metastore with
>>>>> URI thrift://ip-xxx-xx-xx-xxx.ap-southeast-1.compute.internal:9083
>>>>> 16/02/29 23:09:30 INFO metastore: Connected to metastore.
>>>>> 16/02/29 23:09:30 WARN DomainSocketFactory: The short-circuit local
>>>>> reads feature cannot be used because libhadoop cannot be loaded.
>>>>> 16/02/29 23:09:31 INFO SessionState: Created local directory:
>>>>> /tmp/1bf53785-f7c8-406d-a733-a5858ccb2d16_resources
>>>>> 16/02/29 23:09:31 INFO SessionState: Created HDFS directory:
>>>>> /tmp/hive/hdfs/1bf53785-f7c8-406d-a733-a5858ccb2d16
>>>>> 16/02/29 23:09:31 INFO SessionState: Created local directory:
>>>>> /tmp/hdfs/1bf53785-f7c8-406d-a733-a5858ccb2d16
>>>>> 16/02/29 23:09:31 INFO SessionState: Created HDFS directory:
>>>>> /tmp/hive/hdfs/1bf53785-f7c8-406d-a733-a5858ccb2d16/_tmp_space.db
>>>>> hiveContext: org.apache.spark.sql.hive.HiveContext =
>>>>> org.apache.spark.sql.hive.HiveContext@10b14f32
>>>>> 16/02/29 23:09:32 INFO ParseDriver: Parsing command: from TEST SELECT
>>>>> NAME
>>>>> 16/02/29 23:09:32 INFO ParseDriver: Parse Completed
>>>>> 16/02/29 23:09:33 INFO deprecation: mapred.map.tasks is deprecated.
>>>>> Instead, use mapreduce.job.maps
>>>>> 16/02/29 23:09:33 INFO MemoryStore: ensureFreeSpace(468352) called with
>>>>> curMem=0, maxMem=556038881
>>>>> 16/02/29 23:09:33 INFO MemoryStore: Block broadcast_0 stored as values
>>>>> in memory (estimated size 457.4 KB, free 529.8 MB)
>>>>> 16/02/29 23:09:33 INFO MemoryStore: ensureFreeSpace(49454) called with
>>>>> curMem=468352, maxMem=556038881
>>>>> 16/02/29 23:09:33 INFO MemoryStore: Block broadcast_0_piece0 stored as
>>>>> bytes in memory (estimated size 48.3 KB, free 529.8 MB)
>>>>> 16/02/29 23:09:33 INFO BlockManagerInfo: Added broadcast_0_piece0 in
>>>>> memory on xxx.xx.xx.xxx:37784 (size: 48.3 KB, free: 530.2 MB)
>>>>> 16/02/29 23:09:33 INFO SparkContext: Created broadcast 0 from collect
>>>>> at <console>:30
>>>>> 16/02/29 23:09:34 INFO HBaseStorageHandler: Configuring input job
>>>>> properties
>>>>> 16/02/29 23:09:34 INFO RecoverableZooKeeper: Process
>>>>> identifier=hconnection-0x26fa89a2 connecting to ZooKeeper
>>>>> ensemble=localhost:2181
>>>>> 16/02/29 23:09:34 INFO ZooKeeper: Client
>>>>> environment:zookeeper.version=3.4.6-3485--1, built on 12/16/2015 02:35 GMT
>>>>> 16/02/29 23:09:34 INFO ZooKeeper: Client
>>>>> environment:host.name=ip-xxx-xx-xx-xxx.ap-southeast-1.compute.internal
>>>>> 16/02/29 23:09:34 INFO ZooKeeper: Client
>>>>> environment:java.version=1.7.0_67
>>>>> 16/02/29 23:09:34 INFO ZooKeeper: Client environment:java.vendor=Oracle
>>>>> Corporation
>>>>> 16/02/29 23:09:34 INFO ZooKeeper: Client
>>>>> environment:java.home=/usr/jdk64/jdk1.7.0_67/jre
>>>>> 16/02/29 23:09:34 INFO ZooKeeper: Client
>>>>> environment:java.class.path=/usr/hdp/2.3.4.0-3485/hive/lib/guava-14.0.1.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-client.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-common.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-protocol.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-server.jar,/usr/hdp/2.3.4.0-3485/hive/lib/hive-hbase-handler.jar,/usr/hdp/2.3.4.0-3485/hive/lib/htrace-core-3.1.0-incubating.jar,/usr/hdp/2.3.4.0-3485/hive/lib/zookeeper-3.4.6.2.3.4.0-3485.jar,/usr/hdp/2.3.4.0-3485/hive/lib/hive-hbase-handler.jar,/usr/hdp/2.3.4.0-3485/hbase/lib/hbase-server.jar:/usr/hdp/current/spark-thriftserver/conf/:/usr/hdp/2.3.4.0-3485/spark/lib/spark-assembly-1.5.2.2.3.4.0-3485-hadoop2.7.1.2.3.4.0-3485.jar:/usr/hdp/2.3.4.0-3485/spark/lib/datanucleus-api-jdo-3.2.6.jar:/usr/hdp/2.3.4.0-3485/spark/lib/datanucleus-core-3.2.10.jar:/usr/hdp/2.3.4.0-3485/spark/lib/datanucleus-rdbms-3.2.9.jar:/usr/hdp/current/hadoop-client/conf/
>>>>> 16/02/29 23:09:34 INFO ZooKeeper: Client
>>>>> environment:java.library.path=/usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/lib
>>>>> 16/02/29 23:09:34 INFO ZooKeeper: Client
>>>>> environment:java.io.tmpdir=/tmp
>>>>> 16/02/29 23:09:34 INFO ZooKeeper: Client environment:java.compiler=<NA>
>>>>> 16/02/29 23:09:34 INFO ZooKeeper: Client environment:os.name=Linux
>>>>> 16/02/29 23:09:34 INFO ZooKeeper: Client environment:os.arch=amd64
>>>>> 16/02/29 23:09:34 INFO ZooKeeper: Client
>>>>> environment:os.version=3.10.0-229.el7.x86_64
>>>>> 16/02/29 23:09:34 INFO ZooKeeper: Client environment:user.name=hdfs
>>>>> 16/02/29 23:09:34 INFO ZooKeeper: Client
>>>>> environment:user.home=/home/hdfs
>>>>> 16/02/29 23:09:34 INFO ZooKeeper: Client
>>>>> environment:user.dir=/home/hdfs
>>>>> 16/02/29 23:09:34 INFO ZooKeeper: Initiating client connection,
>>>>> connectString=localhost:2181 sessionTimeout=90000
>>>>> watcher=hconnection-0x26fa89a20x0, quorum=localhost:2181, baseZNode=/hbase
>>>>> 16/02/29 23:09:34 INFO ClientCnxn: Opening socket connection to server
>>>>> localhost/0:0:0:0:0:0:0:1:2181. Will not attempt to authenticate using 
>>>>> SASL
>>>>> (unknown error)
>>>>> 16/02/29 23:09:34 INFO ClientCnxn: Socket connection established to
>>>>> localhost/0:0:0:0:0:0:0:1:2181, initiating session
>>>>> 16/02/29 23:09:34 INFO ClientCnxn: Session establishment complete on
>>>>> server localhost/0:0:0:0:0:0:0:1:2181, sessionid = 0x3532fb70ba20034,
>>>>> negotiated timeout = 40000
>>>>> 16/02/29 23:09:34 WARN TableInputFormatBase: You are using an HTable
>>>>> instance that relies on an HBase-managed Connection. This is usually due 
>>>>> to
>>>>> directly creating an HTable, which is deprecated. Instead, you should 
>>>>> create
>>>>> a Connection object and then request a Table instance from it. If you 
>>>>> don't
>>>>> need the Table instance for your own use, you should instead use the
>>>>> TableInputFormatBase.initalizeTable method directly.
>>>>> 16/02/29 23:09:34 INFO TableInputFormatBase: Creating an additional
>>>>> unmanaged connection because user provided one can't be used for
>>>>> administrative actions. We'll close it when we close out the table.
>>>>> 16/02/29 23:09:34 INFO RecoverableZooKeeper: Process
>>>>> identifier=hconnection-0x6fd74d35 connecting to ZooKeeper
>>>>> ensemble=localhost:2181
>>>>> 16/02/29 23:09:34 INFO ZooKeeper: Initiating client connection,
>>>>> connectString=localhost:2181 sessionTimeout=90000
>>>>> watcher=hconnection-0x6fd74d350x0, quorum=localhost:2181, baseZNode=/hbase
>>>>> 16/02/29 23:09:34 INFO ClientCnxn: Opening socket connection to server
>>>>> localhost/0:0:0:0:0:0:0:1:2181. Will not attempt to authenticate using 
>>>>> SASL
>>>>> (unknown error)
>>>>> 16/02/29 23:09:34 INFO ClientCnxn: Socket connection established to
>>>>> localhost/0:0:0:0:0:0:0:1:2181, initiating session
>>>>> 16/02/29 23:09:34 INFO ClientCnxn: Session establishment complete on
>>>>> server localhost/0:0:0:0:0:0:0:1:2181, sessionid = 0x3532fb70ba20035,
>>>>> negotiated timeout = 40000
>>>>> 16/02/29 23:09:34 INFO RegionSizeCalculator: Calculating region sizes
>>>>> for table "TEST".
>>>>> 16/02/29 23:10:23 INFO RpcRetryingCaller: Call exception, tries=10,
>>>>> retries=35, started=48318 ms ago, cancelled=false, msg=
>>>>> 16/02/29 23:10:43 INFO RpcRetryingCaller: Call exception, tries=11,
>>>>> retries=35, started=68524 ms ago, cancelled=false, msg=
>>>>> 16/02/29 23:11:03 INFO RpcRetryingCaller: Call exception, tries=12,
>>>>> retries=35, started=88617 ms ago, cancelled=false, msg=
>>>>> 16/02/29 23:11:23 INFO RpcRetryingCaller: Call exception, tries=13,
>>>>> retries=35, started=108676 ms ago, cancelled=false, msg=
>>>>> 16/02/29 23:11:43 INFO RpcRetryingCaller: Call exception, tries=14,
>>>>> retries=35, started=128747 ms ago, cancelled=false, msg=
>>>>> 16/02/29 23:12:03 INFO RpcRetryingCaller: Call exception, tries=15,
>>>>> retries=35, started=148938 ms ago, cancelled=false, msg=
>>>>> 16/02/29 23:12:23 INFO RpcRetryingCaller: Call exception, tries=16,
>>>>> retries=35, started=168942 ms ago, cancelled=false, msg=
>>>>> 16/02/29 23:12:43 INFO RpcRetryingCaller: Call exception, tries=17,
>>>>> retries=35, started=188975 ms ago, cancelled=false, msg=
>>>>> Trace :
>>>>
>>>>
>>>>
>>>> Could somebody help me in resolving the error.
>>>> Would really appreciate the help .
>>>>
>>>>
>>>> Thanks,
>>>> Divya
>>>>
>>>>
>>>>
>>>>
>>>>
>>>
>>
>

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to