Github user dongjoon-hyun commented on the issue:

    https://github.com/apache/spark/pull/20846
  
    @liutang123 . Did you test this with the latest Apache Spark 2.3? Apache 
Spark 2.3 works without any problem with your example.
    ```scala
    scala> sql("create table test_par(a string) PARTITIONED BY (b bigint) ROW 
FORMAT SERDE 'org.apache.hadoop.hive.ql.io.orc.OrcSerde' STORED AS INPUTFORMAT 
'org.apache.hadoop.hive.ql.io.orc.OrcInputFormat' OUTPUTFORMAT 
'org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat'")
    res0: org.apache.spark.sql.DataFrame = []
    
    hive> ALTER TABLE test_par CHANGE a a bigint restrict;
    OK
    Time taken: 1.358 seconds
    
    scala> sql("select * from test_par").show
    18/03/16 17:33:52 WARN ObjectStore: Failed to get database global_temp, 
returning NoSuchObjectException
    18/03/16 17:33:53 WARN HiveExternalCatalog: The table schema given by Hive 
metastore(struct<a:bigint,b:bigint>) is different from the schema when this 
table was created by Spark SQL(struct<a:string,b:bigint>). We have to fall back 
to the table schema from Hive metastore which is not case preserving.
    18/03/16 17:33:54 WARN HiveExternalCatalog: The table schema given by Hive 
metastore(struct<a:bigint,b:bigint>) is different from the schema when this 
table was created by Spark SQL(struct<a:string,b:bigint>). We have to fall back 
to the table schema from Hive metastore which is not case preserving.
    +---+---+
    |  a|  b|
    +---+---+
    +---+---+
    
    
    scala> sc.version
    res1: String = 2.3.0
    ```
    
    So, please include a test case for this PR. You may insert some data to 
illustrate your issue.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to