thanks,

 hdfs of hive is  :2.7.1
 hdfs of hbase is :2.4.0
 version of hbase is :0.98.12
________________________________
and kylin.hbase.cluster.fs is set to 
:kylin.hbase.cluster.fs=hdfs://10.209.21.11:10010

full log:

[pool-5-thread-1]:[2015-11-12 
10:20:07,961][INFO][org.apache.kylin.job.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:57)]
 - parameters of the HadoopShellExecutable:
[pool-5-thread-1]:[2015-11-12 
10:20:07,962][INFO][org.apache.kylin.job.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:58)]
 -  -input 
hdfs://10.209.21.11:10010/tmp/kylin/kylin_metadata/kylin-061e3569-a973-4c0a-8d89-84c9e0b12117/test2/hfile
 -htablename KYLIN_S71CJH8UZB -cubename test2
[pool-5-thread-1]:[2015-11-12 
10:20:08,067][DEBUG][org.apache.kylin.job.hadoop.hbase.BulkLoadJob.run(BulkLoadJob.java:86)]
 - Start to run LoadIncrementalHFiles
[pool-5-thread-1]:[2015-11-12 
10:20:09,108][ERROR][org.apache.kylin.job.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:64)]
 - error execute 
HadoopShellExecutable{id=061e3569-a973-4c0a-8d89-84c9e0b12117-10, name=Load 
HFile to HBase Table, state=RUNNING}
java.io.IOException: BulkLoad encountered an unrecoverable problem
        at 
org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.bulkLoadPhase(LoadIncrementalHFiles.java:443)
        at 
org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.doBulkLoad(LoadIncrementalHFiles.java:375)
        at 
org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.run(LoadIncrementalHFiles.java:951)
        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
        at 
org.apache.kylin.job.hadoop.hbase.BulkLoadJob.run(BulkLoadJob.java:87)
        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
        at 
org.apache.kylin.job.common.HadoopShellExecutable.doWork(HadoopShellExecutable.java:62)
        at 
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
        at 
org.apache.kylin.job.execution.DefaultChainedExecutable.doWork(DefaultChainedExecutable.java:51)
        at 
org.apache.kylin.job.execution.AbstractExecutable.execute(AbstractExecutable.java:107)
        at 
org.apache.kylin.job.impl.threadpool.DefaultScheduler$JobRunner.run(DefaultScheduler.java:130)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed 
after attempts=3, exceptions:
Wed Nov 11 18:20:08 GMT-08:00 2015, 
org.apache.hadoop.hbase.client.RpcRetryingCaller@7544d58e, java.io.IOException: 
java.io.IOException: Wrong FS: 
hdfs://10.209.21.11:10010/tmp/kylin/kylin_metadata/kylin-061e3569-a973-4c0a-8d89-84c9e0b12117/test2/hfile/F1/fae5ec4c764e4bd8921d6a43d2295493,
 expected: hdfs://wanda
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2132)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
        at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.IllegalArgumentException: Wrong FS: 
hdfs://10.209.21.11:10010/tmp/kylin/kylin_metadata/kylin-061e3569-a973-4c0a-8d89-84c9e0b12117/test2/hfile/F1/fae5ec4c764e4bd8921d6a43d2295493,
 expected: hdfs://wanda
        at org.apache.hadoop.fs.FileSystem.checkPath(FileSystem.java:642)
        at 
org.apache.hadoop.hdfs.DistributedFileSystem.getPathName(DistributedFileSystem.java:181)
        at 
org.apache.hadoop.hdfs.DistributedFileSystem.access$000(DistributedFileSystem.java:92)
        at 
org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:1106)
        at 
org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:1102)
        at 
org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
        at 
org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1102)
        at 
org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:398)
        at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1397)
        at 
org.apache.hadoop.hbase.regionserver.HRegionFileSystem.commitStoreFile(HRegionFileSystem.java:367)
        at 
org.apache.hadoop.hbase.regionserver.HRegionFileSystem.bulkLoadStoreFile(HRegionFileSystem.java:446)
        at 
org.apache.hadoop.hbase.regionserver.HStore.bulkLoadHFile(HStore.java:694)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.bulkLoadHFiles(HRegion.java:3855)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.bulkLoadHFiles(HRegion.java:3761)
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.bulkLoadHFile(HRegionServer.java:3426)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:30948)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2093)
        ... 4 more

Wed Nov 11 18:20:08 GMT-08:00 2015, 
org.apache.hadoop.hbase.client.RpcRetryingCaller@7544d58e, java.io.IOException: 
java.io.IOException: Wrong FS: 
hdfs://10.209.21.11:10010/tmp/kylin/kylin_metadata/kylin-061e3569-a973-4c0a-8d89-84c9e0b12117/test2/hfile/F1/fae5ec4c764e4bd8921d6a43d2295493,
 expected: hdfs://wanda
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2132)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
        at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.IllegalArgumentException: Wrong FS: 
hdfs://10.209.21.11:10010/tmp/kylin/kylin_metadata/kylin-061e3569-a973-4c0a-8d89-84c9e0b12117/test2/hfile/F1/fae5ec4c764e4bd8921d6a43d2295493,
 expected: hdfs://wanda
        at org.apache.hadoop.fs.FileSystem.checkPath(FileSystem.java:642)
        at 
org.apache.hadoop.hdfs.DistributedFileSystem.getPathName(DistributedFileSystem.java:181)
        at 
org.apache.hadoop.hdfs.DistributedFileSystem.access$000(DistributedFileSystem.java:92)
        at 
org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:1106)
        at 
org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:1102)
        at 
org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
        at 
org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1102)
        at 
org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:398)
        at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1397)
        at 
org.apache.hadoop.hbase.regionserver.HRegionFileSystem.commitStoreFile(HRegionFileSystem.java:367)
        at 
org.apache.hadoop.hbase.regionserver.HRegionFileSystem.bulkLoadStoreFile(HRegionFileSystem.java:446)
        at 
org.apache.hadoop.hbase.regionserver.HStore.bulkLoadHFile(HStore.java:694)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.bulkLoadHFiles(HRegion.java:3855)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.bulkLoadHFiles(HRegion.java:3761)
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.bulkLoadHFile(HRegionServer.java:3426)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:30948)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2093)
        ... 4 more


Which Kylin version you're using? Besides, seems the stack trace you pasted
is truncked, please provide the fulll error stack so we can see it was
triggered from which part of Kylin.

Also, did you configure "kylin.hbase.cluster.fs" in "conf/kylin.properties"
with your HBase HDFS host info?

2015-11-11 21:02 GMT+08:00 Xiaoyu Wang <[hidden 
email]<http://apache-kylin-incubating.74782.x6.nabble.com/user/SendEmail.jtp?type=node&node=2312&i=0>>:

> Which version Hadoop,HBase do you use?
>
>
> 在 2015年11月11日 20:49, LIU Ze (刘则) 写道:
>
>> hi,all
>>
>> in the step of "Load HFile to HBase Table"  ,connect to regionserver will
>> retry 35 times .
>>
>> the regionserver hosts had add to /etc/hosts   ,why it can not connect to
>> regionserver?
>> ________________________________
>>
>> 2015-11-11 20:32:04,549 DEBUG [LoadIncrementalHFiles-1]
>> mapreduce.LoadIncrementalHFiles: Going to connect to server
>> region=KYLIN_A1CZUZI0MU,,1447244994197.ae3549d9a76c212fea1330f849d4b4ae.,
>> hostname=CDM1C22-209021018,10620,1444379096591, seqNum=1 for row  with
>> hfile group [{[B@760ed940,hdfs://
>> 10.209.21.11:10010/tmp/kylin/kylin_metadata/kylin-dd2f8f9f-76b1-44bb-8a7e-96b9c3924e0a/test2/hfile/F1/b4507237baf944d2818933935f22441d
>> }]
>> 2015-11-11 20:32:04,558 INFO  [LoadIncrementalHFiles-1]
>> client.RpcRetryingCaller: Call exception, tries=12, retries=35,
>> retryTime=108455ms, msg=row '' on table 'KYLIN_A1CZUZI0MU' at
>> region=KYLIN_A1CZUZI0MU,,1447244994197.ae3549d9a76c212fea1330f849d4b4ae.,
>> hostname=CDM1C22-209021018,10620,1444379096591, seqNum=1
>> 2015-11-11 20:32:24,577 DEBUG [LoadIncrementalHFiles-1]
>> client.HConnectionManager$HConnectionImplementation: Removed
>> CDM1C22-209021018:10620 as a location of
>> KYLIN_A1CZUZI0MU,,1447244994197.ae3549d9a76c212fea1330f849d4b4ae. for
>> tableName=KYLIN_A1CZUZI0MU from cache
>>
>> hdfs of hbase:hdfs://10.209.21.11:10010
>> hdfs of hive:  hdfs://wanda
>>
>> Wed Nov 11 04:30:16 GMT-08:00 2015,
>> org.apache.hadoop.hbase.client.RpcRetryingCaller@53cb23f2,
>> java.io.IOException: java.io.IOException: Wrong FS: hdfs://
>> 10.209.21.11:10010/tmp/kylin/kylin_metadata/kylin-dd2f8f9f-76b1-44bb-8a7e-96b9c3924
>> e0a/test2/hfile/F1/b4507237baf944d2818933935f22441d, expected:
>> hdfs://wanda
>>          at
>> org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2132)
>>          at
>> org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
>>          at
>> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
>>          at
>> org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
>>          at java.lang.Thread.run(Thread.java:745)
>> Caused by: java.lang.IllegalArgumentException: Wrong FS: hdfs://
>> 10.209.21.11:10010/tmp/kylin/kylin_metadata/kylin-dd2f8f9f-76b1-44bb-8a7e-96b9c3924e0a/test2/hfile/F1/b4507237baf944d2818933935f22441d,
>> expected: hdfs://wanda
>>          at org.apache.hadoop.fs.FileSystem.checkPath(FileSystem.java:642)
>>          at
>> org.apache.hadoop.hdfs.DistributedFileSystem.getPathName(DistributedFileSystem.java:181)
>>          at
>> org.apache.hadoop.hdfs.DistributedFileSystem.access$000(DistributedFileSystem.java:92)
>>          at
>> org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:1106)
>>          at
>> org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:1102)
>>
>
>
... [show rest of quote<javascript:void(0)>]


--
Best regards,

Shaofeng Shi

Reply via email to