[ 
https://issues.apache.org/jira/browse/CARBONDATA-3926?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17165566#comment-17165566
 ] 

yutao commented on CARBONDATA-3926:
-----------------------------------

but i think  it can  be a hdfs directory 

> flink-integration i find it can't move file to stage_data directory 
> --------------------------------------------------------------------
>
>                 Key: CARBONDATA-3926
>                 URL: https://issues.apache.org/jira/browse/CARBONDATA-3926
>             Project: CarbonData
>          Issue Type: Bug
>          Components: flink-integration
>    Affects Versions: 2.0.0, 2.0.1
>         Environment: my hadoop is cdh-5.16.1 and spark 2.3.3, flink 
> 1.10.1,hive 1.1.0
>            Reporter: yutao
>            Priority: Critical
>             Fix For: 2.1.0
>
>
> [https://github.com/apache/carbondata/blob/master/docs/flink-integration-guide.md]
>  i work with this ,use spark sql create carbondata table and i can see 
>  -rw-r--r-- 3 hadoop dc_cbss 2650 2020-07-25 21:06 
> hdfs://beh/user/dc_cbss/warehouse/testyu.db/userpolicy/Metadata/schema
> then i write flink app and run with yarn;
> it work i can see carbonfile in my code defined directory ;
> val dataTempPath = "hdfs://beh/user/dc_cbss/temp/"
> [dc_cbss@hive_client_004 yutao]$ hdfs dfs -ls hdfs://beh/user/dc_cbss/temp/
>  Found 10 items
>  drwxr-xr-x - dc_cbss dc_cbss 0 2020-07-27 14:47 
> hdfs://beh/user/dc_cbss/temp/33976d2f23344768b91c6ba3eadd22c8
>  drwxr-xr-x - dc_cbss dc_cbss 0 2020-07-27 14:35 
> hdfs://beh/user/dc_cbss/temp/359a873ec9624623af9beae18b630fde
>  drwxr-xr-x - dc_cbss dc_cbss 0 2020-07-27 14:44 
> hdfs://beh/user/dc_cbss/temp/372f6065515e41a5b1d5e01af0a78d61
>  drwxr-xr-x - dc_cbss dc_cbss 0 2020-07-27 14:50 
> hdfs://beh/user/dc_cbss/temp/3735b94780484f96b211ff6d6974ce3a
>  drwxr-xr-x - dc_cbss dc_cbss 0 2020-07-27 14:38 
> hdfs://beh/user/dc_cbss/temp/8411793f4c5547dc930aacaeea3177cd
>  drwxr-xr-x - dc_cbss dc_cbss 0 2020-07-27 14:29 
> hdfs://beh/user/dc_cbss/temp/915ff23f0d9e4c2dab699d1dcc5a8b4e
>  drwxr-xr-x - dc_cbss dc_cbss 0 2020-07-27 14:32 
> hdfs://beh/user/dc_cbss/temp/bea0bef07d5f47cd92541c69b16aa64e
>  drwxr-xr-x - dc_cbss dc_cbss 0 2020-07-27 14:26 
> hdfs://beh/user/dc_cbss/temp/c42c760144da4f9d83104af270ed46c1
>  drwxr-xr-x - dc_cbss dc_cbss 0 2020-07-27 14:41 
> hdfs://beh/user/dc_cbss/temp/d8af69e47a5844a3a8ed7090ea13a278
>  drwxr-xr-x - dc_cbss dc_cbss 0 2020-07-27 14:50 
> hdfs://beh/user/dc_cbss/temp/db6dceb913444c92a3453903fb50f486
>  [dc_cbss@hive_client_004 yutao]$ hdfs dfs -ls 
> hdfs://beh/user/dc_cbss/temp/33976d2f23344768b91c6ba3eadd22c8/
>  Found 8 items
>  -rw-r--r-- 3 dc_cbss dc_cbss 3100 2020-07-27 14:45 
> hdfs://beh/user/dc_cbss/temp/33976d2f23344768b91c6ba3eadd22c8/24b93d2ffbc14472b3c0e3d2cd948632_batchno0-0-null-1595831979508.carbonindex
>  -rw-r--r-- 3 dc_cbss dc_cbss 3104 2020-07-27 14:47 
> hdfs://beh/user/dc_cbss/temp/33976d2f23344768b91c6ba3eadd22c8/2da284a3beed4c15a3b60c7849d2da92_batchno0-0-null-1595832075416.carbonindex
>  -rw-r--r-- 3 dc_cbss dc_cbss 3104 2020-07-27 14:47 
> hdfs://beh/user/dc_cbss/temp/33976d2f23344768b91c6ba3eadd22c8/70b01854c2d446889b91d4bc9203587c_batchno0-0-null-1595832123015.carbonindex
>  -rw-r--r-- 3 dc_cbss dc_cbss 3110 2020-07-27 14:46 
> hdfs://beh/user/dc_cbss/temp/33976d2f23344768b91c6ba3eadd22c8/aae80851ef534c9ca6f95669d56ec636_batchno0-0-null-1595832028966.carbonindex
>  -rw-r--r-- 3 dc_cbss dc_cbss 54526 2020-07-27 14:45 
> hdfs://beh/user/dc_cbss/temp/33976d2f23344768b91c6ba3eadd22c8/part-0-24b93d2ffbc14472b3c0e3d2cd948632_batchno0-0-null-1595831979508.snappy.carbondata
>  -rw-r--r-- 3 dc_cbss dc_cbss 54710 2020-07-27 14:47 
> hdfs://beh/user/dc_cbss/temp/33976d2f23344768b91c6ba3eadd22c8/part-0-2da284a3beed4c15a3b60c7849d2da92_batchno0-0-null-1595832075416.snappy.carbondata
>  -rw-r--r-- 3 dc_cbss dc_cbss 38684 2020-07-27 14:47 
> hdfs://beh/user/dc_cbss/temp/33976d2f23344768b91c6ba3eadd22c8/part-0-70b01854c2d446889b91d4bc9203587c_batchno0-0-null-1595832123015.snappy.carbondata
>  -rw-r--r-- 3 dc_cbss dc_cbss 55229 2020-07-27 14:46 
> hdfs://beh/user/dc_cbss/temp/33976d2f23344768b91c6ba3eadd22c8/part-0-aae80851ef534c9ca6f95669d56ec636_batchno0-0-null-1595832028966.snappy.carbondata
>  
> but there no stage_data directory and data not mv to stage_data when flink 
> app commit;
> i debug code find in CarbonWriter.java file find  this method influence it ;
> protected StageInput uploadSegmentDataFiles(final String localPath, final 
> String remotePath) {
> if (!this.table.isHivePartitionTable()) {
>  final *{color:#ff0000}File[] files = new File(localPath).listFiles();{color}*
>  if (files == null)
> { LOGGER.error("files is null" ); return null; }
> Map<String, Long> fileNameMapLength = new HashMap<>(files.length);
>  for (File file : files) {
>  fileNameMapLength.put(file.getName(), file.length());
>  if (LOGGER.isDebugEnabled())
> { LOGGER.debug( "Upload file[" + file.getAbsolutePath() + "] to [" + 
> remotePath + "] start."); }
> try
> { CarbonUtil.copyCarbonDataFileToCarbonStorePath(file.getAbsolutePath(), 
> remotePath, 1024); }
> catch (CarbonDataWriterException exception)
> { LOGGER.error(exception.getMessage(), exception); throw exception; }
> if (LOGGER.isDebugEnabled())
> { LOGGER.debug("Upload file[" + file.getAbsolutePath() + "] to [" + 
> remotePath + "] end."); }
> }
>  return new StageInput(remotePath, fileNameMapLength);
>  } else {
>  final List<StageInput.PartitionLocation> partitionLocationList = new 
> ArrayList<>();
>  final List<String> partitions = new ArrayList<>();
>  uploadSegmentDataFiles(new File(localPath), remotePath, 
> partitionLocationList, partitions);
>  if (partitionLocationList.isEmpty())
> { return null; }
> else
> { return new StageInput(remotePath, partitionLocationList); }
> }
>  the local path is a hdfs file so {color:#ff0000}files is null ;{color}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to