Github user dilipbiswal commented on a diff in the pull request:
https://github.com/apache/spark/pull/9652#discussion_r44891704
--- Diff:
sql/core/src/main/scala/org/apache/spark/sql/sources/interfaces.scala ---
@@ -440,9 +441,16 @@ abstract class HadoopFsRelation
private[sql](maybePartitionSpec: Option[Partitio
val hdfsPath = new Path(path)
val fs = hdfsPath.getFileSystem(hadoopConf)
val qualified = hdfsPath.makeQualified(fs.getUri,
fs.getWorkingDirectory)
-
- logInfo(s"Listing $qualified on driver")
- Try(fs.listStatus(qualified)).getOrElse(Array.empty)
+ logInfo(s"Listing $qualified on dr iver")
+ // Dummy jobconf to get to the pathFilter defined in
configuration
+ val jobConf = new JobConf(hadoopConf, this.getClass())
+ val pathFilter = FileInputFormat.getInputPathFilter(jobConf)
+ if (pathFilter != null) {
+ Try(fs.listStatus(qualified,
pathFilter)).getOrElse(Array.empty)
--- End diff --
@yhuai Thanks for taking a look and your comments. Yin, actually i am not
sure about the usage of Try. Is it because fs.listStatus can potentially throw
an exception and subsequent call to getOrElse ? I just followed the existing
logic .. Please let me know what you think ..
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]