zhedoubushishi commented on a change in pull request #2208:
URL: https://github.com/apache/hudi/pull/2208#discussion_r533108221



##########
File path: 
hudi-spark/src/main/scala/org/apache/hudi/MergeOnReadSnapshotRelation.scala
##########
@@ -113,9 +113,6 @@ class MergeOnReadSnapshotRelation(val sqlContext: 
SQLContext,
       hadoopConf = sqlContext.sparkSession.sessionState.newHadoopConf()
     )
 
-    // Follow the implementation of Spark internal HadoopRDD to handle the 
broadcast configuration.

Review comment:
       ```SparkHadoopUtil``` becomes ```private``` in spark 3, so I need to see 
if these lines are necessary.
   
   The reason this is needed in internal implementation is to guard against 
case where user passes a custom Configuration which doesn't contain credentials 
to access secure HDFS (https://github.com/apache/spark/pull/2676). Since the 
Configuration being used here was created as part of spark context the 
credentials should already be loaded. So we can remove it.




----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
[email protected]


Reply via email to