I'm now hitting a document issue. Trying to apply the fix suggested in the
JIRA

https://issues.apache.org/jira/browse/KYLIN-953

when cube job run at the "Convert Cuboid Data to HFile" step, throws an
error like bellow:
[pool-5-thread-8]:[2015-08-18 09:43:15,854][ERROR]
[org.apache.kylin.job.hadoop.cube.CubeHFileJob.run(CubeHFileJob.java:98)] -
error in CubeHFileJ
ob
java.lang.IllegalArgumentException: Can not create a Path from a null string
at org.apache.hadoop.fs.Path.checkPathArg(Path.java:123)

On Tue, Oct 20, 2015 at 10:49 AM, sdangi <[email protected]> wrote:

> Hi Luke -- I will be looking into this later today.  But here is the
> progress (or lack there of), so far
>
> 1)   cd /home/worker1/kylin/1.x-HBase1.x
>
> 2)   [root@worker1 1.x-HBase1.x]# git clone -b 1.x-HBase1.x
> https://github.com/apache/incubator-kylin.git .
>
> [root@worker1 1.x-HBase1.x]# ls -ltr
>
> total 88
>
> -rw-r--r--  1 root root   849 Oct 20 10:13 README.md
>
> -rw-r--r--  1 root root   180 Oct 20 10:13 NOTICE
>
> -rw-r--r--  1 root root 12401 Oct 20 10:13 LICENSE
>
> -rw-r--r--  1 root root  7290 Oct 20 10:13 KEYS
>
> -rw-r--r--  1 root root   539 Oct 20 10:13 DISCLAIMER
>
> drwxr-xr-x  4 root root    46 Oct 20 10:13 atopcalcite
>
> drwxr-xr-x  4 root root    46 Oct 20 10:13 common
>
> drwxr-xr-x  2 root root  4096 Oct 20 10:13 bin
>
> drwxr-xr-x  2 root root    54 Oct 20 10:13 conf
>
> drwxr-xr-x  4 root root    46 Oct 20 10:13 cube
>
> drwxr-xr-x  4 root root    46 Oct 20 10:13 dictionary
>
> drwxr-xr-x  2 root root    23 Oct 20 10:13 deploy
>
> drwxr-xr-x  2 root root    22 Oct 20 10:13 docs
>
> drwxr-xr-x  4 root root    62 Oct 20 10:13 examples
>
> drwxr-xr-x  4 root root    46 Oct 20 10:13 invertedindex
>
> drwxr-xr-x  4 root root    46 Oct 20 10:13 jdbc
>
> drwxr-xr-x  4 root root    96 Oct 20 10:13 job
>
> drwxr-xr-x  4 root root    46 Oct 20 10:13 metadata
>
> drwxr-xr-x  4 root root    46 Oct 20 10:13 monitor
>
> drwxr-xr-x  4 root root    46 Oct 20 10:13 query
>
> -rw-r--r--  1 root root 39837 Oct 20 10:13 pom.xml
>
> drwxr-xr-x  2 root root    98 Oct 20 10:13 script
>
> drwxr-xr-x  4 root root    69 Oct 20 10:13 server
>
> drwxr-xr-x  3 root root    17 Oct 20 10:13 src
>
> drwxr-xr-x  4 root root    46 Oct 20 10:13 storage
>
> drwxr-xr-x  3 root root  4096 Oct 20 10:13 webapp
>
> drwxr-xr-x 16 root root  4096 Oct 20 10:13 website
>
>
>
>
>
>
>
> Build usingMaven
>
> INFO] --- maven-assembly-plugin:2.5.5:single (make-assembly) @
> kylin-monitor ---
>
> [WARNING] Artifact:
> org.apache.kylin:kylin-monitor:jar:1.1-incubating-SNAPSHOT references the
> same file as the assembly destination file. Moving it to a temporary
> location for inclusion.
>
> [INFO] Building jar:
>
> /home/worker1/kylin/1.x-HBase1.x/monitor/target/kylin-monitor-1.1-incubating-SNAPSHOT.jar
>
> [WARNING] Configuration options: 'appendAssemblyId' is set to false, and
> 'classifier' is missing.
>
> Instead of attaching the assembly file:
>
> /home/worker1/kylin/1.x-HBase1.x/monitor/target/kylin-monitor-1.1-incubating-SNAPSHOT.jar,
> it will become the file for main project artifact.
>
> NOTE: If multiple descriptors or descriptor-formats are provided for this
> project, the value of this file will be non-deterministic!
>
> [WARNING] Replacing pre-existing project main-artifact file:
>
> /home/worker1/kylin/1.x-HBase1.x/monitor/target/archive-tmp/kylin-monitor-1.1-incubating-SNAPSHOT.jar
>
> with assembly file:
>
> /home/worker1/kylin/1.x-HBase1.x/monitor/target/kylin-monitor-1.1-incubating-SNAPSHOT.jar
>
> [INFO]
>
> [INFO] --- maven-jar-plugin:2.4:test-jar (default) @ kylin-monitor ---
>
> [INFO] Building jar:
>
> /home/worker1/kylin/1.x-HBase1.x/monitor/target/kylin-monitor-1.1-incubating-SNAPSHOT-tests.jar
>
> [INFO]
> ------------------------------------------------------------------------
>
> [INFO] Reactor Summary:
>
> [INFO]
>
> [INFO] Kylin:HadoopOLAPEngine ............................. SUCCESS [
> 0.864 s]
>
> [INFO] Kylin:AtopCalcite .................................. SUCCESS [
> 5.439 s]
>
> [INFO] Kylin:Common ....................................... SUCCESS [
> 7.231 s]
>
> [INFO] Kylin:Metadata ..................................... SUCCESS [
> 1.428 s]
>
> [INFO] Kylin:Dictionary ................................... SUCCESS [
> 1.559 s]
>
> [INFO] Kylin:Cube ......................................... SUCCESS [
> 2.344 s]
>
> [INFO] Kylin:InvertedIndex ................................ SUCCESS [
> 0.523 s]
>
> [INFO] Kylin:Job .......................................... SUCCESS [
> 3.889 s]
>
> [INFO] Kylin:Storage ...................................... SUCCESS [
> 2.018 s]
>
> [INFO] Kylin:Query ........................................ SUCCESS [
> 1.278 s]
>
> [INFO] Kylin:JDBC ......................................... SUCCESS [
> 1.901 s]
>
> [INFO] Kylin:RESTServer ................................... SUCCESS [
> 8.819 s]
>
> [INFO] Kylin:Monitor ...................................... SUCCESS [
> 1.038 s]
>
> [INFO]
> ------------------------------------------------------------------------
>
> [INFO] BUILD SUCCESS
>
> [INFO]
> ------------------------------------------------------------------------
>
> [INFO] Total time: 38.658 s
>
> [INFO] Finished at: 2015-10-20T10:17:54-04:00
>
> [INFO] Final Memory: 132M/2053M
>
> [INFO]
> ------------------------------------------------------------------------
>
>
>
>
>
>  Imported the sample cube and ran job.  It goes up to the 13th step and
> fails build Hbase tables - seems like a permission issue.
>
>
> ==> kylin.log <==
>
> [pool-7-thread-10]:[2015-10-20
>
> 10:44:58,078][INFO][org.apache.kylin.job.tools.DeployCoprocessorCLI.deployCoprocessor(DeployCoprocessorCLI.java:99)]
> - hbase table [B@a092af6 deployed with coprocessor.
>
> usage: CreateHTableJob
>
>  -cubename <name>            Cube name. For exmaple, flat_item_cube
>
>  -htablename <htable name>   HTable name
>
>  -input <path>               Partition file path.
>
> org.apache.hadoop.hbase.DoNotRetryIOException:
> org.apache.hadoop.hbase.DoNotRetryIOException: Mkdirs failed to create
> /tmp/hbase-hbase/local/jars/tmp (exists=false, cwd=file:/home/hbase) Set
> hbase.table.sanity.checks to false at conf or table descriptor if you want
> to bypass sanity checks
>
> at
>
> org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1597)
>
> at
>
> org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1529)
>
> at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1448)
>
> at
>
> org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:422)
>
> at
>
> org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:48502)
>
> at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
>
> at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
>
> at
> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
>
> at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
>
> at java.lang.Thread.run(Thread.java:745)
>
>
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
>
> at
>
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
>
> at
>
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>
> at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
>
> at
>
> org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
>
> at
>
> org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:95)
>
> at
>
> org.apache.hadoop.hbase.client.RpcRetryingCaller.translateException(RpcRetryingCaller.java:226)
>
> at
>
> org.apache.hadoop.hbase.client.RpcRetryingCaller.translateException(RpcRetryingCaller.java:240)
>
> at
>
> org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:140)
>
> at
>
> org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:3917)
>
> at
>
> org.apache.hadoop.hbase.client.HBaseAdmin.createTableAsyncV2(HBaseAdmin.java:636)
>
> at
> org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:557)
>
> at
>
> org.apache.kylin.job.hadoop.hbase.CreateHTableJob.run(CreateHTableJob.java:150)
>
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
>
> at
>
> org.apache.kylin.job.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:62)
>
> at
>
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
>
> at
>
> org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:51)
>
> at
>
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
>
> at
>
> org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:130)
>
> at
>
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>
> at
>
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>
> at java.lang.Thread.run(Thread.java:745)
>
> Caused by:
>
> org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.DoNotRetryIOException):
> org.apache.hadoop.hbase.DoNotRetryIOException: Mkdirs failed to create
> /tmp/hbase-hbase/local/jars/tmp (exists=false, cwd=file:/home/hbase) Set
> hbase.table.sanity.checks to false at conf or table descriptor if you want
> to bypass sanity checks
>
> at
>
> org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1597)
>
> at
>
> org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1529)
>
> at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1448)
>
> at
>
> org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:422)
>
> at
>
> org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:48502)
>
> at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
>
> at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
>
> at
> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
>
> at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
>
> at java.lang.Thread.run(Thread.java:745)
>
>
> at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1226)
>
> at
>
> org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
>
> at
>
> org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
>
> at
>
> org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.createTable(MasterProtos.java:51086)
>
> at
>
> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$4.createTable(ConnectionManager.java:1802)
>
> at org.apache.hadoop.hbase.client.HBaseAdmin$4.call(HBaseAdmin.java:641)
>
> at org.apache.hadoop.hbase.client.HBaseAdmin$4.call(HBaseAdmin.java:637)
>
> at
>
> org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:126)
>
> ... 14 more
>
>
> ==> kylin_job.log <==
>
> [pool-7-thread-10]:[2015-10-20
>
> 10:44:58,113][ERROR][org.apache.kylin.job.hadoop.hbase.CreateHTableJob.run(CreateHTableJob.java:157)]
> - org.apache.hadoop.hbase.DoNotRetryIOException: Mkdirs failed to create
> /tmp/hbase-hbase/local/jars/tmp (exists=false, cwd=file:/home/hbase) Set
> hbase.table.sanity.checks to false at conf or table descriptor if you want
> to bypass sanity checks
>
> at
>
> org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1597)
>
> at
>
> org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1529)
>
> at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1448)
>
> at
>
> org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:422)
>
> at
>
> org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:48502)
>
> at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
>
> at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
>
> at
> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
>
> at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
>
> at java.lang.Thread.run(Thread.java:745)
>
>
> org.apache.hadoop.hbase.DoNotRetryIOException:
> org.apache.hadoop.hbase.DoNotRetryIOException: Mkdirs failed to create
> /tmp/hbase-hbase/local/jars/tmp (exists=false, cwd=file:/home/hbase) Set
> hbase.table.sanity.checks to false at conf or table descriptor if you want
> to bypass sanity checks
>
> at
>
> org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1597)
>
> at
>
> org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1529)
>
> at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1448)
>
> at
>
> org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:422)
>
> at
>
> org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:48502)
>
> at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
>
> at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
>
> at
> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
>
> at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
>
> at java.lang.Thread.run(Thread.java:745)
>
>
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
>
> at
>
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
>
> at
>
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>
> at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
>
> at
>
> org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
>
> at
>
> org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:95)
>
> at
>
> org.apache.hadoop.hbase.client.RpcRetryingCaller.translateException(RpcRetryingCaller.java:226)
>
> at
>
> org.apache.hadoop.hbase.client.RpcRetryingCaller.translateException(RpcRetryingCaller.java:240)
>
> at
>
> org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:140)
>
> at
>
> org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:3917)
>
> at
>
> org.apache.hadoop.hbase.client.HBaseAdmin.createTableAsyncV2(HBaseAdmin.java:636)
>
> at
> org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:557)
>
> at
>
> org.apache.kylin.job.hadoop.hbase.CreateHTableJob.run(CreateHTableJob.java:150)
>
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
>
> at
>
> org.apache.kylin.job.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:62)
>
> at
>
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
>
> at
>
> org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:51)
>
> at
>
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
>
> at
>
> org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:130)
>
> at
>
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>
> at
>
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>
> at java.lang.Thread.run(Thread.java:745)
>
> Caused by:
>
> org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.DoNotRetryIOException):
> org.apache.hadoop.hbase.DoNotRetryIOException: Mkdirs failed to create
> /tmp/hbase-hbase/local/jars/tmp (exists=false, cwd=file:/home/hbase) Set
> hbase.table.sanity.checks to false at conf or table descriptor if you want
> to bypass sanity checks
>
> at
>
> org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1597)
>
> at
>
> org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1529)
>
> at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1448)
>
> at
>
> org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:422)
>
> at
>
> org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:48502)
>
> at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
>
> at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
>
> at
> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
>
> at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
>
> at java.lang.Thread.run(Thread.java:745)
>
>
> at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1226)
>
> at
>
> org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
>
> at
>
> org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
>
> at
>
> org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.createTable(MasterProtos.java:51086)
>
> at
>
> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$4.createTable(ConnectionManager.java:1802)
>
> at org.apache.hadoop.hbase.client.HBaseAdmin$4.call(HBaseAdmin.java:641)
>
> at org.apache.hadoop.hbase.client.HBaseAdmin$4.call(HBaseAdmin.java:637)
>
> at
>
> org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:126)
>
> ... 14 more
>
>
> ==> kylin.log <==
>
> [pool-7-thread-10]:[2015-10-20
>
> 10:44:58,113][ERROR][org.apache.kylin.job.hadoop.hbase.CreateHTableJob.run(CreateHTableJob.java:157)]
> - org.apache.hadoop.hbase.DoNotRetryIOException: Mkdirs failed to create
> /tmp/hbase-hbase/local/jars/tmp (exists=false, cwd=file:/home/hbase) Set
> hbase.table.sanity.checks to false at conf or table descriptor if you want
> to bypass sanity checks
>
> at
>
> org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1597)
>
> at
>
> org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1529)
>
> at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1448)
>
> at
>
> org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:422)
>
> at
>
> org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:48502)
>
> at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
>
> at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
>
> at
> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
>
> at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
>
> at java.lang.Thread.run(Thread.java:745)
>
>
> org.apache.hadoop.hbase.DoNotRetryIOException:
> org.apache.hadoop.hbase.DoNotRetryIOException: Mkdirs failed to create
> /tmp/hbase-hbase/local/jars/tmp (exists=false, cwd=file:/home/hbase) Set
> hbase.table.sanity.checks to false at conf or table descriptor if you want
> to bypass sanity checks
>
> at
>
> org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1597)
>
> at
>
> org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1529)
>
> at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1448)
>
> at
>
> org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:422)
>
> at
>
> org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:48502)
>
> at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
>
> at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
>
> at
> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
>
> at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
>
> at java.lang.Thread.run(Thread.java:745)
>
>
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
>
> at
>
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
>
> at
>
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>
> at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
>
> at
>
> org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
>
> at
>
> org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:95)
>
> at
>
> org.apache.hadoop.hbase.client.RpcRetryingCaller.translateException(RpcRetryingCaller.java:226)
>
> at
>
> org.apache.hadoop.hbase.client.RpcRetryingCaller.translateException(RpcRetryingCaller.java:240)
>
> at
>
> org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:140)
>
> at
>
> org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:3917)
>
> at
>
> org.apache.hadoop.hbase.client.HBaseAdmin.createTableAsyncV2(HBaseAdmin.java:636)
>
> at
> org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:557)
>
> at
>
> org.apache.kylin.job.hadoop.hbase.CreateHTableJob.run(CreateHTableJob.java:150)
>
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>
> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
>
> at
>
> org.apache.kylin.job.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:62)
>
> at
>
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
>
> at
>
> org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:51)
>
> at
>
> org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
>
> at
>
> org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:130)
>
> at
>
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>
> at
>
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>
> at java.lang.Thread.run(Thread.java:745)
>
> Caused by:
>
> org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.DoNotRetryIOException):
> org.apache.hadoop.hbase.DoNotRetryIOException: Mkdirs failed to create
> /tmp/hbase-hbase/local/jars/tmp (exists=false, cwd=file:/home/hbase) Set
> hbase.table.sanity.checks to false at conf or table descriptor if you want
> to bypass sanity checks
>
> at
>
> org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1597)
>
> at
>
> org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1529)
>
> at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1448)
>
> at
>
> org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:422)
>
> at
>
> org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:48502)
>
> at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
>
> at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
>
> at
> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
>
> at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
>
> at java.lang.Thread.run(Thread.java:745)
>
>
> at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1226)
>
> at
>
> org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
>
> at
>
> org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
>
> at
>
> org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.createTable(MasterProtos.java:51086)
>
> at
>
> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$4.createTable(ConnectionManager.java:1802)
>
> at org.apache.hadoop.hbase.client.HBaseAdmin$4.call(HBaseAdmin.java:641)
>
> at org.apache.hadoop.hbase.client.HBaseAdmin$4.call(HBaseAdmin.java:637)
>
> at
>
> org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:126)
>
> ... 14 more
>
> 2015-10-20 10:44:58,115 INFO  [pool-7-thread-10]
> client.ConnectionManager$HConnectionImplementation: Closing master
> protocol: MasterService
>
> 2015-10-20 10:44:58,115 INFO  [pool-7-thread-10]
> client.ConnectionManager$HConnectionImplementation: Closing zookeeper
> sessionid=0x15067463b2f001b
>
> 2015-10-20 10:44:58,145 INFO  [pool-7-thread-10] zookeeper.ZooKeeper:
> Session: 0x15067463b2f001b closed
>
> 2015-10-20 10:44:58,145 INFO  [pool-7-thread-10-EventThread]
> zookeeper.ClientCnxn: EventThread shut down
>
> [pool-7-thread-10]:[2015-10-20
>
> 10:44:58,161][DEBUG][org.apache.kylin.common.persistence.ResourceStore.putResource(ResourceStore.java:195)]
> - Saving resource /execute_output/476fa1ea-25ea-4858-b780-51028c298274-12
> (Store kylin_metadata@hbase)
>
> [pool-7-thread-10]:[2015-10-20
>
> 10:44:58,172][DEBUG][org.apache.kylin.common.persistence.ResourceStore.putResource(ResourceStore.java:195)]
> - Saving resource /execute_output/476fa1ea-25ea-4858-b780-51028c298274-12
> (Store kylin_metadata@hbase)
>
> On Sat, Oct 17, 2015 at 11:13 AM, Luke Han [via Apache Kylin (Incubating)]
> <
> [email protected]> wrote:
>
> > Hi Shailesh,
> >     If timing is concern, we strongly suggest to downgrade your HBase to
> > 0.98 with Kylin. the 1.x branch is not fully tested yet.
> >
> >     If you still would like to try with HBase 1.x, please clone this
> > branch:
> >     https://github.com/apache/incubator-kylin/tree/1.x-HBase1.x
> >
> >     And, then run ./script/package.sh to generate binary package
> >     Then copy package from dist folder and install with your Hadoop
> > cluster.
> >
> >      BTW, which distribution you are using now? CDH or HDP?
> >
> >     Thanks.
> >
> > Luke
> >
> >
> > Best Regards!
> > ---------------------
> >
> > Luke Han
> >
> > On Sat, Oct 17, 2015 at 8:29 AM, sdangi <[hidden email]
> > <http:///user/SendEmail.jtp?type=node&node=1996&i=0>> wrote:
> >
> > > Luke/Kylin Team -- Any further updates/guidance you could offer?
> Latest
> > > clone does not work w/ 1.1 version of HBase.
> > >
> > > We are working on a time sensitive POC for a financial client and
> > > appreciate
> > > your responses.
> > >
> > > Thanks,
> > > Regards,
> > >
> > >
> > >
> > > --
> > > View this message in context:
> > >
> >
> http://apache-kylin-incubating.74782.x6.nabble.com/SAMPLE-CUBE-FAILS-tp1936p1994.html
> > > Sent from the Apache Kylin (Incubating) mailing list archive at
> > Nabble.com.
> > >
> >
> >
> > ------------------------------
> > If you reply to this email, your message will be added to the discussion
> > below:
> >
> >
> http://apache-kylin-incubating.74782.x6.nabble.com/SAMPLE-CUBE-FAILS-tp1936p1996.html
> > To unsubscribe from SAMPLE CUBE FAILS, click here
> > <
> http://apache-kylin-incubating.74782.x6.nabble.com/template/NamlServlet.jtp?macro=unsubscribe_by_code&node=1936&code=c2RhbmdpQGRhdGFsZW56LmNvbXwxOTM2fDQzMTE2MjM5NA==
> >
> > .
> > NAML
> > <
> http://apache-kylin-incubating.74782.x6.nabble.com/template/NamlServlet.jtp?macro=macro_viewer&id=instant_html%21nabble%3Aemail.naml&base=nabble.naml.namespaces.BasicNamespace-nabble.view.web.template.NabbleNamespace-nabble.naml.namespaces.BasicNamespace-nabble.view.web.template.NabbleNamespace-nabble.view.web.template.NodeNamespace&breadcrumbs=notify_subscribers%21nabble%3Aemail.naml-instant_emails%21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml
> >
> >
>
>
>
>
> --
> View this message in context:
> http://apache-kylin-incubating.74782.x6.nabble.com/SAMPLE-CUBE-FAILS-tp1936p2037.html
> Sent from the Apache Kylin (Incubating) mailing list archive at Nabble.com.
>

Reply via email to