Github user ppadma commented on the issue: https://github.com/apache/drill/pull/1030 @arina-ielchiieva I am concerned about performance impact by grouping all splits in a single reader (essentially, not parallelizing at all). Wondering if it is possible to do this way: During planning, in HiveScan, if it is text file and has header/footer, get the number of rows to skip. Read the header/footer rows and based on that, adjust the first/last split and offset within them. The splits which have only header/footer rows can be removed from inputSplits. In HiveSubScan, change hiveReadEntry to be a list (one entry for each split). Add an entry in hiveReadEntry, numRowsToSkip (or offsetToStart) which can be passed to the recordReaders in getBatch for each subScan. This is fairly complicated and I am sure I might be missing some details :-)
---