Yash Datta created SPARK-6742:
---------------------------------

             Summary: Spark pushes down filters in old parquet path that 
reference partitioning columns
                 Key: SPARK-6742
                 URL: https://issues.apache.org/jira/browse/SPARK-6742
             Project: Spark
          Issue Type: Bug
          Components: SQL
    Affects Versions: 1.2.1
            Reporter: Yash Datta


Create a table with multiple fields partitioned on 'market' column. run a query 
like : 

SELECT start_sp_time, end_sp_time, imsi, imei,  enb_common_enbid FROM 
csl_data_parquet WHERE (((technology = 'FDD') AND (bandclass = '800') AND 
(region = 'R15') AND (market = 'LA metro')) OR ((technology = 'FDD') AND 
(bandclass = '1900') AND (region = 'R15') AND (market = 'Indianapolis'))) AND 
start_sp_time >= 1.4158368E9 AND end_sp_time < 1.4159232E9 AND dt >= 
'2014-11-13-00-00' AND dt < '2014-11-14-00-00' ORDER BY end_sp_time DESC LIMIT 
100

The or filter is pushed down in this case , resulting in column not found 
exception from parquet 




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to