Github user dongjoon-hyun commented on a diff in the pull request:
https://github.com/apache/spark/pull/19651#discussion_r148932599
--- Diff:
sql/hive/src/main/scala/org/apache/spark/sql/hive/orc/OrcFileOperator.scala ---
@@ -87,15 +88,10 @@ private[hive] object OrcFileOperator extends Logging {
getFileReader(path,
conf).map(_.getObjectInspector.asInstanceOf[StructObjectInspector])
}
- def listOrcFiles(pathStr: String, conf: Configuration): Seq[Path] = {
- // TODO: Check if the paths coming in are already qualified and
simplify.
- val origPath = new Path(pathStr)
- val fs = origPath.getFileSystem(conf)
- val paths = SparkHadoopUtil.get.listLeafStatuses(fs, origPath)
- .filterNot(_.isDirectory)
- .map(_.getPath)
- .filterNot(_.getName.startsWith("_"))
- .filterNot(_.getName.startsWith("."))
- paths
+ def setRequiredColumns(
--- End diff --
This is moved from [object
ORCFileFormat](https://github.com/apache/spark/pull/19651/files#diff-01999ccbf13e95a0ea2d223f69d8ae23L265)
inside `sql/hive`.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]