To sum up what's in https://issues.apache.org/jira/browse/KYLIN-953. To resolve the issue, you need one of the two hadoop config below in your site xmls
- hadoop.tmp.dir (for hbase 1.1.0 and before) - hbase.fs.tmp.dir (for hbase 1.1.1 and after) On Wed, Oct 21, 2015 at 1:07 AM, Shailesh Dangi <[email protected]> wrote: > I'm now hitting a document issue. Trying to apply the fix suggested in the > JIRA > > https://issues.apache.org/jira/browse/KYLIN-953 > > when cube job run at the "Convert Cuboid Data to HFile" step, throws an > error like bellow: > [pool-5-thread-8]:[2015-08-18 09:43:15,854][ERROR] > [org.apache.kylin.job.hadoop.cube.CubeHFileJob.run(CubeHFileJob.java:98)] - > error in CubeHFileJ > ob > java.lang.IllegalArgumentException: Can not create a Path from a null > string > at org.apache.hadoop.fs.Path.checkPathArg(Path.java:123) > > On Tue, Oct 20, 2015 at 10:49 AM, sdangi <[email protected]> wrote: > > > Hi Luke -- I will be looking into this later today. But here is the > > progress (or lack there of), so far > > > > 1) cd /home/worker1/kylin/1.x-HBase1.x > > > > 2) [root@worker1 1.x-HBase1.x]# git clone -b 1.x-HBase1.x > > https://github.com/apache/incubator-kylin.git . > > > > [root@worker1 1.x-HBase1.x]# ls -ltr > > > > total 88 > > > > -rw-r--r-- 1 root root 849 Oct 20 10:13 README.md > > > > -rw-r--r-- 1 root root 180 Oct 20 10:13 NOTICE > > > > -rw-r--r-- 1 root root 12401 Oct 20 10:13 LICENSE > > > > -rw-r--r-- 1 root root 7290 Oct 20 10:13 KEYS > > > > -rw-r--r-- 1 root root 539 Oct 20 10:13 DISCLAIMER > > > > drwxr-xr-x 4 root root 46 Oct 20 10:13 atopcalcite > > > > drwxr-xr-x 4 root root 46 Oct 20 10:13 common > > > > drwxr-xr-x 2 root root 4096 Oct 20 10:13 bin > > > > drwxr-xr-x 2 root root 54 Oct 20 10:13 conf > > > > drwxr-xr-x 4 root root 46 Oct 20 10:13 cube > > > > drwxr-xr-x 4 root root 46 Oct 20 10:13 dictionary > > > > drwxr-xr-x 2 root root 23 Oct 20 10:13 deploy > > > > drwxr-xr-x 2 root root 22 Oct 20 10:13 docs > > > > drwxr-xr-x 4 root root 62 Oct 20 10:13 examples > > > > drwxr-xr-x 4 root root 46 Oct 20 10:13 invertedindex > > > > drwxr-xr-x 4 root root 46 Oct 20 10:13 jdbc > > > > drwxr-xr-x 4 root root 96 Oct 20 10:13 job > > > > drwxr-xr-x 4 root root 46 Oct 20 10:13 metadata > > > > drwxr-xr-x 4 root root 46 Oct 20 10:13 monitor > > > > drwxr-xr-x 4 root root 46 Oct 20 10:13 query > > > > -rw-r--r-- 1 root root 39837 Oct 20 10:13 pom.xml > > > > drwxr-xr-x 2 root root 98 Oct 20 10:13 script > > > > drwxr-xr-x 4 root root 69 Oct 20 10:13 server > > > > drwxr-xr-x 3 root root 17 Oct 20 10:13 src > > > > drwxr-xr-x 4 root root 46 Oct 20 10:13 storage > > > > drwxr-xr-x 3 root root 4096 Oct 20 10:13 webapp > > > > drwxr-xr-x 16 root root 4096 Oct 20 10:13 website > > > > > > > > > > > > > > > > Build usingMaven > > > > INFO] --- maven-assembly-plugin:2.5.5:single (make-assembly) @ > > kylin-monitor --- > > > > [WARNING] Artifact: > > org.apache.kylin:kylin-monitor:jar:1.1-incubating-SNAPSHOT references the > > same file as the assembly destination file. Moving it to a temporary > > location for inclusion. > > > > [INFO] Building jar: > > > > > /home/worker1/kylin/1.x-HBase1.x/monitor/target/kylin-monitor-1.1-incubating-SNAPSHOT.jar > > > > [WARNING] Configuration options: 'appendAssemblyId' is set to false, and > > 'classifier' is missing. > > > > Instead of attaching the assembly file: > > > > > /home/worker1/kylin/1.x-HBase1.x/monitor/target/kylin-monitor-1.1-incubating-SNAPSHOT.jar, > > it will become the file for main project artifact. > > > > NOTE: If multiple descriptors or descriptor-formats are provided for this > > project, the value of this file will be non-deterministic! > > > > [WARNING] Replacing pre-existing project main-artifact file: > > > > > /home/worker1/kylin/1.x-HBase1.x/monitor/target/archive-tmp/kylin-monitor-1.1-incubating-SNAPSHOT.jar > > > > with assembly file: > > > > > /home/worker1/kylin/1.x-HBase1.x/monitor/target/kylin-monitor-1.1-incubating-SNAPSHOT.jar > > > > [INFO] > > > > [INFO] --- maven-jar-plugin:2.4:test-jar (default) @ kylin-monitor --- > > > > [INFO] Building jar: > > > > > /home/worker1/kylin/1.x-HBase1.x/monitor/target/kylin-monitor-1.1-incubating-SNAPSHOT-tests.jar > > > > [INFO] > > ------------------------------------------------------------------------ > > > > [INFO] Reactor Summary: > > > > [INFO] > > > > [INFO] Kylin:HadoopOLAPEngine ............................. SUCCESS [ > > 0.864 s] > > > > [INFO] Kylin:AtopCalcite .................................. SUCCESS [ > > 5.439 s] > > > > [INFO] Kylin:Common ....................................... SUCCESS [ > > 7.231 s] > > > > [INFO] Kylin:Metadata ..................................... SUCCESS [ > > 1.428 s] > > > > [INFO] Kylin:Dictionary ................................... SUCCESS [ > > 1.559 s] > > > > [INFO] Kylin:Cube ......................................... SUCCESS [ > > 2.344 s] > > > > [INFO] Kylin:InvertedIndex ................................ SUCCESS [ > > 0.523 s] > > > > [INFO] Kylin:Job .......................................... SUCCESS [ > > 3.889 s] > > > > [INFO] Kylin:Storage ...................................... SUCCESS [ > > 2.018 s] > > > > [INFO] Kylin:Query ........................................ SUCCESS [ > > 1.278 s] > > > > [INFO] Kylin:JDBC ......................................... SUCCESS [ > > 1.901 s] > > > > [INFO] Kylin:RESTServer ................................... SUCCESS [ > > 8.819 s] > > > > [INFO] Kylin:Monitor ...................................... SUCCESS [ > > 1.038 s] > > > > [INFO] > > ------------------------------------------------------------------------ > > > > [INFO] BUILD SUCCESS > > > > [INFO] > > ------------------------------------------------------------------------ > > > > [INFO] Total time: 38.658 s > > > > [INFO] Finished at: 2015-10-20T10:17:54-04:00 > > > > [INFO] Final Memory: 132M/2053M > > > > [INFO] > > ------------------------------------------------------------------------ > > > > > > > > > > > > Imported the sample cube and ran job. It goes up to the 13th step and > > fails build Hbase tables - seems like a permission issue. > > > > > > ==> kylin.log <== > > > > [pool-7-thread-10]:[2015-10-20 > > > > > 10:44:58,078][INFO][org.apache.kylin.job.tools.DeployCoprocessorCLI.deployCoprocessor(DeployCoprocessorCLI.java:99)] > > - hbase table [B@a092af6 deployed with coprocessor. > > > > usage: CreateHTableJob > > > > -cubename <name> Cube name. For exmaple, flat_item_cube > > > > -htablename <htable name> HTable name > > > > -input <path> Partition file path. > > > > org.apache.hadoop.hbase.DoNotRetryIOException: > > org.apache.hadoop.hbase.DoNotRetryIOException: Mkdirs failed to create > > /tmp/hbase-hbase/local/jars/tmp (exists=false, cwd=file:/home/hbase) Set > > hbase.table.sanity.checks to false at conf or table descriptor if you > want > > to bypass sanity checks > > > > at > > > > > org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1597) > > > > at > > > > > org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1529) > > > > at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1448) > > > > at > > > > > org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:422) > > > > at > > > > > org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:48502) > > > > at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114) > > > > at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101) > > > > at > > > org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130) > > > > at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107) > > > > at java.lang.Thread.run(Thread.java:745) > > > > > > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > > > > at > > > > > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) > > > > at > > > > > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) > > > > at java.lang.reflect.Constructor.newInstance(Constructor.java:422) > > > > at > > > > > org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106) > > > > at > > > > > org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:95) > > > > at > > > > > org.apache.hadoop.hbase.client.RpcRetryingCaller.translateException(RpcRetryingCaller.java:226) > > > > at > > > > > org.apache.hadoop.hbase.client.RpcRetryingCaller.translateException(RpcRetryingCaller.java:240) > > > > at > > > > > org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:140) > > > > at > > > > > org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:3917) > > > > at > > > > > org.apache.hadoop.hbase.client.HBaseAdmin.createTableAsyncV2(HBaseAdmin.java:636) > > > > at > > > org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:557) > > > > at > > > > > org.apache.kylin.job.hadoop.hbase.CreateHTableJob.run(CreateHTableJob.java:150) > > > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) > > > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84) > > > > at > > > > > org.apache.kylin.job.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:62) > > > > at > > > > > org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107) > > > > at > > > > > org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:51) > > > > at > > > > > org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107) > > > > at > > > > > org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:130) > > > > at > > > > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > > > > at > > > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > > > > at java.lang.Thread.run(Thread.java:745) > > > > Caused by: > > > > > org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.DoNotRetryIOException): > > org.apache.hadoop.hbase.DoNotRetryIOException: Mkdirs failed to create > > /tmp/hbase-hbase/local/jars/tmp (exists=false, cwd=file:/home/hbase) Set > > hbase.table.sanity.checks to false at conf or table descriptor if you > want > > to bypass sanity checks > > > > at > > > > > org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1597) > > > > at > > > > > org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1529) > > > > at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1448) > > > > at > > > > > org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:422) > > > > at > > > > > org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:48502) > > > > at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114) > > > > at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101) > > > > at > > > org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130) > > > > at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107) > > > > at java.lang.Thread.run(Thread.java:745) > > > > > > at > org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1226) > > > > at > > > > > org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213) > > > > at > > > > > org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287) > > > > at > > > > > org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.createTable(MasterProtos.java:51086) > > > > at > > > > > org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$4.createTable(ConnectionManager.java:1802) > > > > at org.apache.hadoop.hbase.client.HBaseAdmin$4.call(HBaseAdmin.java:641) > > > > at org.apache.hadoop.hbase.client.HBaseAdmin$4.call(HBaseAdmin.java:637) > > > > at > > > > > org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:126) > > > > ... 14 more > > > > > > ==> kylin_job.log <== > > > > [pool-7-thread-10]:[2015-10-20 > > > > > 10:44:58,113][ERROR][org.apache.kylin.job.hadoop.hbase.CreateHTableJob.run(CreateHTableJob.java:157)] > > - org.apache.hadoop.hbase.DoNotRetryIOException: Mkdirs failed to create > > /tmp/hbase-hbase/local/jars/tmp (exists=false, cwd=file:/home/hbase) Set > > hbase.table.sanity.checks to false at conf or table descriptor if you > want > > to bypass sanity checks > > > > at > > > > > org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1597) > > > > at > > > > > org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1529) > > > > at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1448) > > > > at > > > > > org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:422) > > > > at > > > > > org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:48502) > > > > at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114) > > > > at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101) > > > > at > > > org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130) > > > > at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107) > > > > at java.lang.Thread.run(Thread.java:745) > > > > > > org.apache.hadoop.hbase.DoNotRetryIOException: > > org.apache.hadoop.hbase.DoNotRetryIOException: Mkdirs failed to create > > /tmp/hbase-hbase/local/jars/tmp (exists=false, cwd=file:/home/hbase) Set > > hbase.table.sanity.checks to false at conf or table descriptor if you > want > > to bypass sanity checks > > > > at > > > > > org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1597) > > > > at > > > > > org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1529) > > > > at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1448) > > > > at > > > > > org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:422) > > > > at > > > > > org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:48502) > > > > at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114) > > > > at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101) > > > > at > > > org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130) > > > > at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107) > > > > at java.lang.Thread.run(Thread.java:745) > > > > > > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > > > > at > > > > > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) > > > > at > > > > > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) > > > > at java.lang.reflect.Constructor.newInstance(Constructor.java:422) > > > > at > > > > > org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106) > > > > at > > > > > org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:95) > > > > at > > > > > org.apache.hadoop.hbase.client.RpcRetryingCaller.translateException(RpcRetryingCaller.java:226) > > > > at > > > > > org.apache.hadoop.hbase.client.RpcRetryingCaller.translateException(RpcRetryingCaller.java:240) > > > > at > > > > > org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:140) > > > > at > > > > > org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:3917) > > > > at > > > > > org.apache.hadoop.hbase.client.HBaseAdmin.createTableAsyncV2(HBaseAdmin.java:636) > > > > at > > > org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:557) > > > > at > > > > > org.apache.kylin.job.hadoop.hbase.CreateHTableJob.run(CreateHTableJob.java:150) > > > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) > > > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84) > > > > at > > > > > org.apache.kylin.job.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:62) > > > > at > > > > > org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107) > > > > at > > > > > org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:51) > > > > at > > > > > org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107) > > > > at > > > > > org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:130) > > > > at > > > > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > > > > at > > > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > > > > at java.lang.Thread.run(Thread.java:745) > > > > Caused by: > > > > > org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.DoNotRetryIOException): > > org.apache.hadoop.hbase.DoNotRetryIOException: Mkdirs failed to create > > /tmp/hbase-hbase/local/jars/tmp (exists=false, cwd=file:/home/hbase) Set > > hbase.table.sanity.checks to false at conf or table descriptor if you > want > > to bypass sanity checks > > > > at > > > > > org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1597) > > > > at > > > > > org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1529) > > > > at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1448) > > > > at > > > > > org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:422) > > > > at > > > > > org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:48502) > > > > at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114) > > > > at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101) > > > > at > > > org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130) > > > > at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107) > > > > at java.lang.Thread.run(Thread.java:745) > > > > > > at > org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1226) > > > > at > > > > > org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213) > > > > at > > > > > org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287) > > > > at > > > > > org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.createTable(MasterProtos.java:51086) > > > > at > > > > > org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$4.createTable(ConnectionManager.java:1802) > > > > at org.apache.hadoop.hbase.client.HBaseAdmin$4.call(HBaseAdmin.java:641) > > > > at org.apache.hadoop.hbase.client.HBaseAdmin$4.call(HBaseAdmin.java:637) > > > > at > > > > > org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:126) > > > > ... 14 more > > > > > > ==> kylin.log <== > > > > [pool-7-thread-10]:[2015-10-20 > > > > > 10:44:58,113][ERROR][org.apache.kylin.job.hadoop.hbase.CreateHTableJob.run(CreateHTableJob.java:157)] > > - org.apache.hadoop.hbase.DoNotRetryIOException: Mkdirs failed to create > > /tmp/hbase-hbase/local/jars/tmp (exists=false, cwd=file:/home/hbase) Set > > hbase.table.sanity.checks to false at conf or table descriptor if you > want > > to bypass sanity checks > > > > at > > > > > org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1597) > > > > at > > > > > org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1529) > > > > at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1448) > > > > at > > > > > org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:422) > > > > at > > > > > org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:48502) > > > > at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114) > > > > at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101) > > > > at > > > org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130) > > > > at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107) > > > > at java.lang.Thread.run(Thread.java:745) > > > > > > org.apache.hadoop.hbase.DoNotRetryIOException: > > org.apache.hadoop.hbase.DoNotRetryIOException: Mkdirs failed to create > > /tmp/hbase-hbase/local/jars/tmp (exists=false, cwd=file:/home/hbase) Set > > hbase.table.sanity.checks to false at conf or table descriptor if you > want > > to bypass sanity checks > > > > at > > > > > org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1597) > > > > at > > > > > org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1529) > > > > at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1448) > > > > at > > > > > org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:422) > > > > at > > > > > org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:48502) > > > > at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114) > > > > at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101) > > > > at > > > org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130) > > > > at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107) > > > > at java.lang.Thread.run(Thread.java:745) > > > > > > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > > > > at > > > > > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) > > > > at > > > > > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) > > > > at java.lang.reflect.Constructor.newInstance(Constructor.java:422) > > > > at > > > > > org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106) > > > > at > > > > > org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:95) > > > > at > > > > > org.apache.hadoop.hbase.client.RpcRetryingCaller.translateException(RpcRetryingCaller.java:226) > > > > at > > > > > org.apache.hadoop.hbase.client.RpcRetryingCaller.translateException(RpcRetryingCaller.java:240) > > > > at > > > > > org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:140) > > > > at > > > > > org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:3917) > > > > at > > > > > org.apache.hadoop.hbase.client.HBaseAdmin.createTableAsyncV2(HBaseAdmin.java:636) > > > > at > > > org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:557) > > > > at > > > > > org.apache.kylin.job.hadoop.hbase.CreateHTableJob.run(CreateHTableJob.java:150) > > > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) > > > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84) > > > > at > > > > > org.apache.kylin.job.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:62) > > > > at > > > > > org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107) > > > > at > > > > > org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:51) > > > > at > > > > > org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107) > > > > at > > > > > org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:130) > > > > at > > > > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > > > > at > > > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > > > > at java.lang.Thread.run(Thread.java:745) > > > > Caused by: > > > > > org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.DoNotRetryIOException): > > org.apache.hadoop.hbase.DoNotRetryIOException: Mkdirs failed to create > > /tmp/hbase-hbase/local/jars/tmp (exists=false, cwd=file:/home/hbase) Set > > hbase.table.sanity.checks to false at conf or table descriptor if you > want > > to bypass sanity checks > > > > at > > > > > org.apache.hadoop.hbase.master.HMaster.warnOrThrowExceptionForFailure(HMaster.java:1597) > > > > at > > > > > org.apache.hadoop.hbase.master.HMaster.sanityCheckTableDescriptor(HMaster.java:1529) > > > > at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1448) > > > > at > > > > > org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:422) > > > > at > > > > > org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:48502) > > > > at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114) > > > > at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101) > > > > at > > > org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130) > > > > at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107) > > > > at java.lang.Thread.run(Thread.java:745) > > > > > > at > org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1226) > > > > at > > > > > org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213) > > > > at > > > > > org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287) > > > > at > > > > > org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.createTable(MasterProtos.java:51086) > > > > at > > > > > org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$4.createTable(ConnectionManager.java:1802) > > > > at org.apache.hadoop.hbase.client.HBaseAdmin$4.call(HBaseAdmin.java:641) > > > > at org.apache.hadoop.hbase.client.HBaseAdmin$4.call(HBaseAdmin.java:637) > > > > at > > > > > org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:126) > > > > ... 14 more > > > > 2015-10-20 10:44:58,115 INFO [pool-7-thread-10] > > client.ConnectionManager$HConnectionImplementation: Closing master > > protocol: MasterService > > > > 2015-10-20 10:44:58,115 INFO [pool-7-thread-10] > > client.ConnectionManager$HConnectionImplementation: Closing zookeeper > > sessionid=0x15067463b2f001b > > > > 2015-10-20 10:44:58,145 INFO [pool-7-thread-10] zookeeper.ZooKeeper: > > Session: 0x15067463b2f001b closed > > > > 2015-10-20 10:44:58,145 INFO [pool-7-thread-10-EventThread] > > zookeeper.ClientCnxn: EventThread shut down > > > > [pool-7-thread-10]:[2015-10-20 > > > > > 10:44:58,161][DEBUG][org.apache.kylin.common.persistence.ResourceStore.putResource(ResourceStore.java:195)] > > - Saving resource /execute_output/476fa1ea-25ea-4858-b780-51028c298274-12 > > (Store kylin_metadata@hbase) > > > > [pool-7-thread-10]:[2015-10-20 > > > > > 10:44:58,172][DEBUG][org.apache.kylin.common.persistence.ResourceStore.putResource(ResourceStore.java:195)] > > - Saving resource /execute_output/476fa1ea-25ea-4858-b780-51028c298274-12 > > (Store kylin_metadata@hbase) > > > > On Sat, Oct 17, 2015 at 11:13 AM, Luke Han [via Apache Kylin > (Incubating)] > > < > > [email protected]> wrote: > > > > > Hi Shailesh, > > > If timing is concern, we strongly suggest to downgrade your HBase > to > > > 0.98 with Kylin. the 1.x branch is not fully tested yet. > > > > > > If you still would like to try with HBase 1.x, please clone this > > > branch: > > > https://github.com/apache/incubator-kylin/tree/1.x-HBase1.x > > > > > > And, then run ./script/package.sh to generate binary package > > > Then copy package from dist folder and install with your Hadoop > > > cluster. > > > > > > BTW, which distribution you are using now? CDH or HDP? > > > > > > Thanks. > > > > > > Luke > > > > > > > > > Best Regards! > > > --------------------- > > > > > > Luke Han > > > > > > On Sat, Oct 17, 2015 at 8:29 AM, sdangi <[hidden email] > > > <http:///user/SendEmail.jtp?type=node&node=1996&i=0>> wrote: > > > > > > > Luke/Kylin Team -- Any further updates/guidance you could offer? > > Latest > > > > clone does not work w/ 1.1 version of HBase. > > > > > > > > We are working on a time sensitive POC for a financial client and > > > > appreciate > > > > your responses. > > > > > > > > Thanks, > > > > Regards, > > > > > > > > > > > > > > > > -- > > > > View this message in context: > > > > > > > > > > http://apache-kylin-incubating.74782.x6.nabble.com/SAMPLE-CUBE-FAILS-tp1936p1994.html > > > > Sent from the Apache Kylin (Incubating) mailing list archive at > > > Nabble.com. > > > > > > > > > > > > > ------------------------------ > > > If you reply to this email, your message will be added to the > discussion > > > below: > > > > > > > > > http://apache-kylin-incubating.74782.x6.nabble.com/SAMPLE-CUBE-FAILS-tp1936p1996.html > > > To unsubscribe from SAMPLE CUBE FAILS, click here > > > < > > > http://apache-kylin-incubating.74782.x6.nabble.com/template/NamlServlet.jtp?macro=unsubscribe_by_code&node=1936&code=c2RhbmdpQGRhdGFsZW56LmNvbXwxOTM2fDQzMTE2MjM5NA== > > > > > > . > > > NAML > > > < > > > http://apache-kylin-incubating.74782.x6.nabble.com/template/NamlServlet.jtp?macro=macro_viewer&id=instant_html%21nabble%3Aemail.naml&base=nabble.naml.namespaces.BasicNamespace-nabble.view.web.template.NabbleNamespace-nabble.naml.namespaces.BasicNamespace-nabble.view.web.template.NabbleNamespace-nabble.view.web.template.NodeNamespace&breadcrumbs=notify_subscribers%21nabble%3Aemail.naml-instant_emails%21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml > > > > > > > > > > > > > > > > -- > > View this message in context: > > > http://apache-kylin-incubating.74782.x6.nabble.com/SAMPLE-CUBE-FAILS-tp1936p2037.html > > Sent from the Apache Kylin (Incubating) mailing list archive at > Nabble.com. > > >
