读hive分区表报的什么错啊,把stacktrace贴一下?

On Wed, May 27, 2020 at 6:08 PM Zhou Zach <[email protected]> wrote:

>
>
> hive partition table:
>
>
> 1CREATE TABLE `dwd.bill`(
> 2  `id` bigint,
> 3  `gid` bigint,
> 4  `count` bigint,
> 5  `price` bigint,
> 6  `srcuid` bigint,
> 7  `srcnickname` string,
> 8  `srcleftmoney` bigint,
> 9  `srcwealth` bigint,
> 10  `srccredit` decimal(10,0),
> 11  `dstnickname` string,
> 12  `dstuid` bigint,
> 13  `familyid` int,
> 14  `dstleftmoney` bigint,
> 15  `dstwealth` bigint,
> 16  `dstcredit` decimal(10,0),
> 17  `addtime` bigint,
> 18  `type` int,
> 19  `getmoney` decimal(10,0),
> 20  `os` int,
> 21  `bak` string,
> 22  `getbonus` decimal(10,0),
> 23  `unionbonus` decimal(10,0))
> 24PARTITIONED BY (
> 25  `year` int,
> 26  `month` int)
> 27ROW FORMAT SERDE
> 28  'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe'
> 29STORED AS INPUTFORMAT
> 30  'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat'
> 31OUTPUTFORMAT
> 32  'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat'
>
>
>
>
> Query:
>
>
> tableEnv.sqlUpdate(
>       """
>         |
>         |INSERT INTO catalog2.dwd.orders
>         |select srcuid, price from catalog2.dwd.bill where year = 2020
>         |
>         |
>         |""".stripMargin)
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> 在 2020-05-27 18:01:19,"Leonard Xu" <[email protected]> 写道:
> >Flink 支持hive分区表的,看你在另外一个邮件里贴了,你能把你的hive表和query在邮件里贴下吗?
> >
> >祝好
> >Leonard Xu
> >
> >> 在 2020年5月27日,17:40,Zhou Zach <[email protected]> 写道:
> >>
> >>
> >>
> >>
> >> 感谢回复,表名前加上Catalog和db前缀可以成功访问了。
> >> 现在遇到个问题,flink 读hive
> 分区表时,如果where子句用分区键,比如year过滤就会报错,用表中其他字段过滤是没问题的,是flink 不支持 hive分区表,还是哪个地方没设置对
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >> 在 2020-05-27 17:33:11,"Leonard Xu" <[email protected]> 写道:
> >>> Hi,
> >>>> 因为一个HiveCatalog只能关联一个库
> >>> 一个Catalog是可以关联到多个db的,不同catalog,不同db中表都可以访问的.
> >>>
> >>> Flink SQL> show catalogs;
> >>> default_catalog
> >>> myhive
> >>> Flink SQL> use catalog myhive;
> >>> Flink SQL> show databases;
> >>> default
> >>> hive_test
> >>> hive_test1
> >>> Flink SQL> select * from hive_test.db2_table union select * from
> myhive.hive_test1.db1_table;
> >>> 2020-05-27 17:25:48,565 INFO  org.apache.hadoop.hive.conf.HiveConf
> >>>
> >>>
> >>>
> >>> 祝好
> >>> Leonard Xu
> >>>
> >>>
> >>>> 在 2020年5月27日,10:55,Zhou Zach <[email protected]> 写道:
> >>>>
> >>>> hi all,
> >>>> Flink sql 的HiveCatalog 是不是不能跨库操作啊,就是一个flink
> sql中join的两个表涉及到两个不同到库,因为一个HiveCatalog只能关联一个库
>


-- 
Best regards!
Rui Li

回复