We are using it via ozzie. We had such issue and we solve it by setting default file permission in hdfs so all users will have access to it. I guess there is more advanced solution but for us it was enough
-----Original Message----- From: Ganesh R [[email protected]] Received: יום שישי, 30 ינו 2015, 9:33 To: [email protected] [[email protected]] Subject: CSV Bulkload in oozie Has anyone tried using csv bulk load tool in oozie? I am trying to use CSVBulkLoader in oozie, it does work in command line with hadoop jar command however when i have it in oozie as java action, it fails to load table at the very end. - The actual mapreduce job shows 100% success exactly reflecting the number of rows to be upserted (matching the input csv file). - have hbase-protocol.jar in the oozie share library. I tried to have this jar as distributed cache but no luck. - alternatively, it works when i covert it to shell action and execute as hadoop jar through oozie. I see FileNotFoundException written in logs over and over again, logs as shown below.. much appreciate your inputs on this. 2015-01-30 01:32:40,223 INFO [phoenix-2-thread-0] org.apache.phoenix.mapreduce.CsvBulkLoadTool: Loading HFiles from /tmp/6cbb4bad-6f26-4546-a831-dd0db13c6d88/OOZIE 2015-01-30 01:32:40,255 WARN [phoenix-2-thread-0] org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles: Skipping non-directory file:/tmp/6cbb4bad-6f26-4546-a831-dd0db13c6d88/OOZIE/_SUCCESS 2015-01-30 01:32:40,314 INFO [LoadIncrementalHFiles-8] org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles: Trying to load hfile=file:/tmp/6cbb4bad-6f26-4546-a831-dd0db13c6d88/OOZIE first=\x0D1000000011001\x001001\x00\x80\x00\x01K=Ix\x00 last=\x0D1000000011001\x001001\x00\x80\x00\x01K=Ix\x00 2015-01-30 01:32:40,315 INFO [LoadIncrementalHFiles-10] org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles: Trying to load hfile=file:/tmp/6cbb4bad-6f26-4546-a831-dd0db13c6d88/OOZIE first=\x101000000011002\x001001\x00\x80\x00\x01K=Ix\x00 last=\x101000000011002\x001003\x00\x80\x00\x01K2\xFC\xC0\x00 2015-01-30 01:41:50,602 ERROR [LoadIncrementalHFiles-13] org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles: Encountered unrecoverable error from region server org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=35, exceptions: Fri Jan 30 01:32:40 EST 2015, org.apache.hadoop.hbase.client.RpcRetryingCaller@549072f7, java.io.FileNotFoundException: java.io.FileNotFoundException: File file:/tmp/6cbb4bad-6f26-4546-a831-dd0db13c6d88/ 2015-01-30 01:41:50,602 ERROR [LoadIncrementalHFiles-13] org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles: Encountered unrecoverable error from region server org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=35, exceptions: Fri Jan 30 01:32:40 EST 2015, org.apache.hadoop.hbase.client.RpcRetryingCaller@549072f7, java.io.FileNotFoundException: java.io.FileNotFoundException: File file:/tmp/6cbb4bad-6f26-4546-a831-dd0db13c6d88/ Information in this e-mail and its attachments is confidential and privileged under the TEOCO confidentiality terms that can be reviewed here<http://www.teoco.com/email-disclaimer>.
