umehrot2 commented on a change in pull request #1924:
URL: https://github.com/apache/hudi/pull/1924#discussion_r466779443
##########
File path:
hudi-client/src/main/java/org/apache/hudi/table/action/bootstrap/BootstrapUtils.java
##########
@@ -41,37 +48,87 @@
* Returns leaf folders with files under a path.
* @param fs File System
* @param basePathStr Base Path to look for leaf folders
- * @param filePathFilter Filters to skip directories/paths
+ * @param jsc Java spark context
* @return list of partition paths with files under them.
* @throws IOException
*/
public static List<Pair<String, List<HoodieFileStatus>>>
getAllLeafFoldersWithFiles(FileSystem fs, String basePathStr,
-
PathFilter filePathFilter) throws IOException {
+ JavaSparkContext jsc) throws IOException {
final Path basePath = new Path(basePathStr);
final Map<Integer, List<String>> levelToPartitions = new HashMap<>();
final Map<String, List<HoodieFileStatus>> partitionToFiles = new
HashMap<>();
- FSUtils.processFiles(fs, basePathStr, (status) -> {
- if (status.isFile() && filePathFilter.accept(status.getPath())) {
- String relativePath = FSUtils.getRelativePartitionPath(basePath,
status.getPath().getParent());
- List<HoodieFileStatus> statusList = partitionToFiles.get(relativePath);
- if (null == statusList) {
- Integer level = (int) relativePath.chars().filter(ch -> ch ==
'/').count();
- List<String> dirs = levelToPartitions.get(level);
- if (null == dirs) {
- dirs = new ArrayList<>();
- levelToPartitions.put(level, dirs);
+ PathFilter filePathFilter = getFilePathFilter();
+ PathFilter metaPathFilter = getExcludeMetaPathFilter();
+
+ FileStatus[] topLevelStatuses = fs.listStatus(new Path(basePathStr));
+ List<String> subDirectories = new ArrayList<>();
+
+ List<Pair<HoodieFileStatus, Pair<Integer, String>>> result = new
ArrayList<>();
Review comment:
Only the outer structure is a bit similar in terms of first listing and
taking action on top level files, and then using `spark context` to perform the
same action on sub-directories in parallel. But the inner logic is different
and values being collected are different.
If we really want to re-use the common outer logic, it would require
exploring extracting out the inner logic into `serializable functions` that
would work fine with `spark context` as well. So, to not over-complicate this
PR I can explore this separately if its okay. I have create a new Jira
https://issues.apache.org/jira/browse/HUDI-1158 where I have listed the two
optimizations we discussed about w.r.t to parallel listing behavior:
- The parallelization should be at leaf partition directory level and not
just at the top directory level
- Extract out common code paths
----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
For queries about this service, please contact Infrastructure at:
[email protected]