[jira] [Resolved] (HIVE-28277) HIVE does not support update operations for ICEBERG of type location_based_table.
[ https://issues.apache.org/jira/browse/HIVE-28277?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] yongzhi.shao resolved HIVE-28277. - Fix Version/s: 4.0.0 Resolution: Won't Fix > HIVE does not support update operations for ICEBERG of type > location_based_table. > - > > Key: HIVE-28277 > URL: https://issues.apache.org/jira/browse/HIVE-28277 > Project: Hive > Issue Type: Bug > Components: Iceberg integration >Affects Versions: 4.0.0 > Environment: ICEBERG:1.5.2 > HIVE 4.0.0 >Reporter: yongzhi.shao >Priority: Major > Fix For: 4.0.0 > > > Currently, when I update the location_based_table using hive, hive > incorrectly empties all data directories and metadata directories. > After the update statement is executed, the iceberg table is corrupted. > > {code:java} > --spark 3.4.1 + iceberg 1.5.2: > CREATE TABLE IF NOT EXISTS datacenter.default.test_data_04 ( > id string,name string > ) > using iceberg > PARTITIONED BY (name) > TBLPROPERTIES > ('read.orc.vectorization.enabled'='true','write.format.default'='orc','write.orc.bloom.filter.columns'='id','write.orc.compression-codec'='zstd','write.metadata.previous-versions-max'='3','write.metadata.delete-after-commit.enabled'='true'); > insert into datacenter.default.test_data_04(id,name) > values('1','a'),('2','b'); > --hive4: > CREATE EXTERNAL TABLE default.test_data_04 > STORED BY 'org.apache.iceberg.mr.hive.HiveIcebergStorageHandler' > LOCATION 'hdfs:///iceberg-catalog/warehouse/default/test_data_04' > TBLPROPERTIES > ('iceberg.catalog'='location_based_table','engine.hive.enabled'='true'); > select id,name from default.test_data_04; --2 row > update test_data_04 set name = 'adasd' where id = '1'; > ERROR: > 2024-05-23T10:26:32,028 ERROR [HiveServer2-Background-Pool: Thread-297] > hive.HiveIcebergStorageHandler: Error while trying to commit job: > job_17061635207991_169536, job_17061635207990_169536, > job_17061635207992_169536, starting rollback changes for table: > default.test_data_04 > org.apache.iceberg.exceptions.NoSuchTableException: Table does not exist at > location: /iceberg-catalog/warehouse/default/test_data_04 > BEFORE UPDATE: > ICEBERG TABLE DIR: > [root@ ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 > Found 2 items > drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 > /iceberg-catalog/warehouse/default/test_data_04/data > drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 > /iceberg-catalog/warehouse/default/test_data_04/metadata > AFTER UPDATE: > ICEBERG TABLE DIR: > [root@XXX ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 > Found 3 items > drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 > /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_1 > drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 > /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_2 > drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 > /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_3 > {code} > > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (HIVE-28277) HIVE does not support update operations for ICEBERG of type location_based_table.
[ https://issues.apache.org/jira/browse/HIVE-28277?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17848817#comment-17848817 ] yongzhi.shao commented on HIVE-28277: - 我更新了代码,问题确实消失了.谢谢 > HIVE does not support update operations for ICEBERG of type > location_based_table. > - > > Key: HIVE-28277 > URL: https://issues.apache.org/jira/browse/HIVE-28277 > Project: Hive > Issue Type: Bug > Components: Iceberg integration >Affects Versions: 4.0.0 > Environment: ICEBERG:1.5.2 > HIVE 4.0.0 >Reporter: yongzhi.shao >Priority: Major > > Currently, when I update the location_based_table using hive, hive > incorrectly empties all data directories and metadata directories. > After the update statement is executed, the iceberg table is corrupted. > > {code:java} > --spark 3.4.1 + iceberg 1.5.2: > CREATE TABLE IF NOT EXISTS datacenter.default.test_data_04 ( > id string,name string > ) > using iceberg > PARTITIONED BY (name) > TBLPROPERTIES > ('read.orc.vectorization.enabled'='true','write.format.default'='orc','write.orc.bloom.filter.columns'='id','write.orc.compression-codec'='zstd','write.metadata.previous-versions-max'='3','write.metadata.delete-after-commit.enabled'='true'); > insert into datacenter.default.test_data_04(id,name) > values('1','a'),('2','b'); > --hive4: > CREATE EXTERNAL TABLE default.test_data_04 > STORED BY 'org.apache.iceberg.mr.hive.HiveIcebergStorageHandler' > LOCATION 'hdfs:///iceberg-catalog/warehouse/default/test_data_04' > TBLPROPERTIES > ('iceberg.catalog'='location_based_table','engine.hive.enabled'='true'); > select id,name from default.test_data_04; --2 row > update test_data_04 set name = 'adasd' where id = '1'; > ERROR: > 2024-05-23T10:26:32,028 ERROR [HiveServer2-Background-Pool: Thread-297] > hive.HiveIcebergStorageHandler: Error while trying to commit job: > job_17061635207991_169536, job_17061635207990_169536, > job_17061635207992_169536, starting rollback changes for table: > default.test_data_04 > org.apache.iceberg.exceptions.NoSuchTableException: Table does not exist at > location: /iceberg-catalog/warehouse/default/test_data_04 > BEFORE UPDATE: > ICEBERG TABLE DIR: > [root@ ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 > Found 2 items > drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 > /iceberg-catalog/warehouse/default/test_data_04/data > drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 > /iceberg-catalog/warehouse/default/test_data_04/metadata > AFTER UPDATE: > ICEBERG TABLE DIR: > [root@XXX ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 > Found 3 items > drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 > /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_1 > drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 > /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_2 > drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 > /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_3 > {code} > > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Comment Edited] (HIVE-28277) HIVE does not support update operations for ICEBERG of type location_based_table.
[ https://issues.apache.org/jira/browse/HIVE-28277?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17848817#comment-17848817 ] yongzhi.shao edited comment on HIVE-28277 at 5/23/24 5:15 AM: -- I've updated the code and the problem has gone away. Thank you, sir. was (Author: lisoda): 我更新了代码,问题确实消失了.谢谢 > HIVE does not support update operations for ICEBERG of type > location_based_table. > - > > Key: HIVE-28277 > URL: https://issues.apache.org/jira/browse/HIVE-28277 > Project: Hive > Issue Type: Bug > Components: Iceberg integration >Affects Versions: 4.0.0 > Environment: ICEBERG:1.5.2 > HIVE 4.0.0 >Reporter: yongzhi.shao >Priority: Major > > Currently, when I update the location_based_table using hive, hive > incorrectly empties all data directories and metadata directories. > After the update statement is executed, the iceberg table is corrupted. > > {code:java} > --spark 3.4.1 + iceberg 1.5.2: > CREATE TABLE IF NOT EXISTS datacenter.default.test_data_04 ( > id string,name string > ) > using iceberg > PARTITIONED BY (name) > TBLPROPERTIES > ('read.orc.vectorization.enabled'='true','write.format.default'='orc','write.orc.bloom.filter.columns'='id','write.orc.compression-codec'='zstd','write.metadata.previous-versions-max'='3','write.metadata.delete-after-commit.enabled'='true'); > insert into datacenter.default.test_data_04(id,name) > values('1','a'),('2','b'); > --hive4: > CREATE EXTERNAL TABLE default.test_data_04 > STORED BY 'org.apache.iceberg.mr.hive.HiveIcebergStorageHandler' > LOCATION 'hdfs:///iceberg-catalog/warehouse/default/test_data_04' > TBLPROPERTIES > ('iceberg.catalog'='location_based_table','engine.hive.enabled'='true'); > select id,name from default.test_data_04; --2 row > update test_data_04 set name = 'adasd' where id = '1'; > ERROR: > 2024-05-23T10:26:32,028 ERROR [HiveServer2-Background-Pool: Thread-297] > hive.HiveIcebergStorageHandler: Error while trying to commit job: > job_17061635207991_169536, job_17061635207990_169536, > job_17061635207992_169536, starting rollback changes for table: > default.test_data_04 > org.apache.iceberg.exceptions.NoSuchTableException: Table does not exist at > location: /iceberg-catalog/warehouse/default/test_data_04 > BEFORE UPDATE: > ICEBERG TABLE DIR: > [root@ ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 > Found 2 items > drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 > /iceberg-catalog/warehouse/default/test_data_04/data > drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 > /iceberg-catalog/warehouse/default/test_data_04/metadata > AFTER UPDATE: > ICEBERG TABLE DIR: > [root@XXX ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 > Found 3 items > drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 > /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_1 > drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 > /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_2 > drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 > /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_3 > {code} > > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (HIVE-28277) HIVE does not support update operations for ICEBERG of type location_based_table.
[ https://issues.apache.org/jira/browse/HIVE-28277?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17848814#comment-17848814 ] Butao Zhang commented on HIVE-28277: I didn't reproduce this issue on hive4/master. maybe some other env problem... > HIVE does not support update operations for ICEBERG of type > location_based_table. > - > > Key: HIVE-28277 > URL: https://issues.apache.org/jira/browse/HIVE-28277 > Project: Hive > Issue Type: Bug > Components: Iceberg integration >Affects Versions: 4.0.0 > Environment: ICEBERG:1.5.2 > HIVE 4.0.0 >Reporter: yongzhi.shao >Priority: Major > > Currently, when I update the location_based_table using hive, hive > incorrectly empties all data directories and metadata directories. > After the update statement is executed, the iceberg table is corrupted. > > {code:java} > --spark 3.4.1 + iceberg 1.5.2: > CREATE TABLE IF NOT EXISTS datacenter.default.test_data_04 ( > id string,name string > ) > using iceberg > PARTITIONED BY (name) > TBLPROPERTIES > ('read.orc.vectorization.enabled'='true','write.format.default'='orc','write.orc.bloom.filter.columns'='id','write.orc.compression-codec'='zstd','write.metadata.previous-versions-max'='3','write.metadata.delete-after-commit.enabled'='true'); > insert into datacenter.default.test_data_04(id,name) > values('1','a'),('2','b'); > --hive4: > CREATE EXTERNAL TABLE default.test_data_04 > STORED BY 'org.apache.iceberg.mr.hive.HiveIcebergStorageHandler' > LOCATION 'hdfs:///iceberg-catalog/warehouse/default/test_data_04' > TBLPROPERTIES > ('iceberg.catalog'='location_based_table','engine.hive.enabled'='true'); > select id,name from default.test_data_04; --2 row > update test_data_04 set name = 'adasd' where id = '1'; > ERROR: > 2024-05-23T10:26:32,028 ERROR [HiveServer2-Background-Pool: Thread-297] > hive.HiveIcebergStorageHandler: Error while trying to commit job: > job_17061635207991_169536, job_17061635207990_169536, > job_17061635207992_169536, starting rollback changes for table: > default.test_data_04 > org.apache.iceberg.exceptions.NoSuchTableException: Table does not exist at > location: /iceberg-catalog/warehouse/default/test_data_04 > BEFORE UPDATE: > ICEBERG TABLE DIR: > [root@ ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 > Found 2 items > drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 > /iceberg-catalog/warehouse/default/test_data_04/data > drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 > /iceberg-catalog/warehouse/default/test_data_04/metadata > AFTER UPDATE: > ICEBERG TABLE DIR: > [root@XXX ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 > Found 3 items > drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 > /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_1 > drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 > /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_2 > drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 > /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_3 > {code} > > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28277) HIVE does not support update operations for ICEBERG of type location_based_table.
[ https://issues.apache.org/jira/browse/HIVE-28277?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] yongzhi.shao updated HIVE-28277: Description: Currently, when I update the location_based_table using hive, hive incorrectly empties all data directories and metadata directories. After the update statement is executed, the iceberg table is corrupted. {code:java} --spark 3.4.1 + iceberg 1.5.2: CREATE TABLE IF NOT EXISTS datacenter.default.test_data_04 ( id string,name string ) using iceberg PARTITIONED BY (name) TBLPROPERTIES ('read.orc.vectorization.enabled'='true','write.format.default'='orc','write.orc.bloom.filter.columns'='id','write.orc.compression-codec'='zstd','write.metadata.previous-versions-max'='3','write.metadata.delete-after-commit.enabled'='true'); insert into datacenter.default.test_data_04(id,name) values('1','a'),('2','b'); --hive4: CREATE EXTERNAL TABLE default.test_data_04 STORED BY 'org.apache.iceberg.mr.hive.HiveIcebergStorageHandler' LOCATION 'hdfs:///iceberg-catalog/warehouse/default/test_data_04' TBLPROPERTIES ('iceberg.catalog'='location_based_table','engine.hive.enabled'='true'); select id,name from default.test_data_04; --2 row update test_data_04 set name = 'adasd' where id = '1'; ERROR: 2024-05-23T10:26:32,028 ERROR [HiveServer2-Background-Pool: Thread-297] hive.HiveIcebergStorageHandler: Error while trying to commit job: job_17061635207991_169536, job_17061635207990_169536, job_17061635207992_169536, starting rollback changes for table: default.test_data_04 org.apache.iceberg.exceptions.NoSuchTableException: Table does not exist at location: /iceberg-catalog/warehouse/default/test_data_04 BEFORE UPDATE: ICEBERG TABLE DIR: [root@ ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 Found 2 items drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 /iceberg-catalog/warehouse/default/test_data_04/data drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 /iceberg-catalog/warehouse/default/test_data_04/metadata AFTER UPDATE: ICEBERG TABLE DIR: [root@XXX ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 Found 3 items drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_1 drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_2 drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_3 {code} was: Currently, when I update the location_based_table using hive, hive incorrectly empties all data directories and metadata directories. After the update statement is executed, the iceberg table is corrupted. {code:java} --spark 3.4.1 + iceberg 1.5.2: CREATE TABLE IF NOT EXISTS datacenter.default.test_data_04 ( id string,name string ) using iceberg PARTITIONED BY (name) TBLPROPERTIES ('read.orc.vectorization.enabled'='true','write.format.default'='orc','write.orc.bloom.filter.columns'='id','write.orc.compression-codec'='zstd','write.metadata.previous-versions-max'='3','write.metadata.delete-after-commit.enabled'='true'); insert into datacenter.default.test_data_04(id,name) values('1','a'),('2','b'); --hive4: CREATE EXTERNAL TABLE default.test_data_04 STORED BY 'org.apache.iceberg.mr.hive.HiveIcebergStorageHandler' LOCATION 'hdfs:///iceberg-catalog/warehouse/default/test_data_04' TBLPROPERTIES ('iceberg.catalog'='location_based_table','engine.hive.enabled'='true'); select distinct id,name from (select id,name from default.test_data_04 limit 10) s1; --2 row update test_data_04 set name = 'adasd' where id = '1'; ERROR: 2024-05-23T10:26:32,028 ERROR [HiveServer2-Background-Pool: Thread-297] hive.HiveIcebergStorageHandler: Error while trying to commit job: job_17061635207991_169536, job_17061635207990_169536, job_17061635207992_169536, starting rollback changes for table: default.test_data_04 org.apache.iceberg.exceptions.NoSuchTableException: Table does not exist at location: /iceberg-catalog/warehouse/default/test_data_04 BEFORE UPDATE: ICEBERG TABLE DIR: [root@ ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 Found 2 items drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 /iceberg-catalog/warehouse/default/test_data_04/data drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 /iceberg-catalog/warehouse/default/test_data_04/metadata AFTER UPDATE: ICEBERG TABLE DIR: [root@XXX ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 Found 3 items drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_1 drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_2 drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 /iceberg-catalog/warehouse/default/test_data_04
[jira] [Updated] (HIVE-28277) HIVE does not support update operations for ICEBERG of type location_based_table.
[ https://issues.apache.org/jira/browse/HIVE-28277?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] yongzhi.shao updated HIVE-28277: Issue Type: Bug (was: Improvement) > HIVE does not support update operations for ICEBERG of type > location_based_table. > - > > Key: HIVE-28277 > URL: https://issues.apache.org/jira/browse/HIVE-28277 > Project: Hive > Issue Type: Bug > Components: Iceberg integration >Affects Versions: 4.0.0 > Environment: ICEBERG:1.5.2 > HIVE 4.0.0 >Reporter: yongzhi.shao >Priority: Major > > Currently, when I update the location_based_table using hive, hive > incorrectly empties all data directories and metadata directories. > After the update statement is executed, the iceberg table is corrupted. > > {code:java} > --spark 3.4.1 + iceberg 1.5.2: > CREATE TABLE IF NOT EXISTS datacenter.default.test_data_04 ( > id string,name string > ) > using iceberg > PARTITIONED BY (name) > TBLPROPERTIES > ('read.orc.vectorization.enabled'='true','write.format.default'='orc','write.orc.bloom.filter.columns'='id','write.orc.compression-codec'='zstd','write.metadata.previous-versions-max'='3','write.metadata.delete-after-commit.enabled'='true'); > insert into datacenter.default.test_data_04(id,name) > values('1','a'),('2','b'); > --hive4: > CREATE EXTERNAL TABLE default.test_data_04 > STORED BY 'org.apache.iceberg.mr.hive.HiveIcebergStorageHandler' > LOCATION 'hdfs:///iceberg-catalog/warehouse/default/test_data_04' > TBLPROPERTIES > ('iceberg.catalog'='location_based_table','engine.hive.enabled'='true'); > select distinct id,name from (select id,name from default.test_data_04 limit > 10) s1; --2 row > update test_data_04 set name = 'adasd' where id = '1'; > ERROR: > 2024-05-23T10:26:32,028 ERROR [HiveServer2-Background-Pool: Thread-297] > hive.HiveIcebergStorageHandler: Error while trying to commit job: > job_17061635207991_169536, job_17061635207990_169536, > job_17061635207992_169536, starting rollback changes for table: > default.test_data_04 > org.apache.iceberg.exceptions.NoSuchTableException: Table does not exist at > location: /iceberg-catalog/warehouse/default/test_data_04 > BEFORE UPDATE: > ICEBERG TABLE DIR: > [root@ ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 > Found 2 items > drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 > /iceberg-catalog/warehouse/default/test_data_04/data > drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 > /iceberg-catalog/warehouse/default/test_data_04/metadata > AFTER UPDATE: > ICEBERG TABLE DIR: > [root@XXX ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 > Found 3 items > drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 > /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_1 > drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 > /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_2 > drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 > /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_3 > {code} > > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28277) HIVE does not support update operations for ICEBERG of type location_based_table.
[ https://issues.apache.org/jira/browse/HIVE-28277?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] yongzhi.shao updated HIVE-28277: Description: Currently, when I update the location_based_table using hive, hive incorrectly empties all data directories and metadata directories. After the update statement is executed, the iceberg table is corrupted. {code:java} --spark 3.4.1 + iceberg 1.5.2: CREATE TABLE IF NOT EXISTS datacenter.default.test_data_04 ( id string,name string ) using iceberg PARTITIONED BY (name) TBLPROPERTIES ('read.orc.vectorization.enabled'='true','write.format.default'='orc','write.orc.bloom.filter.columns'='id','write.orc.compression-codec'='zstd','write.metadata.previous-versions-max'='3','write.metadata.delete-after-commit.enabled'='true'); insert into datacenter.default.test_data_04(id,name) values('1','a'),('2','b'); --hive4: CREATE EXTERNAL TABLE default.test_data_04 STORED BY 'org.apache.iceberg.mr.hive.HiveIcebergStorageHandler' LOCATION 'hdfs:///iceberg-catalog/warehouse/default/test_data_04' TBLPROPERTIES ('iceberg.catalog'='location_based_table','engine.hive.enabled'='true'); select distinct id,name from (select id,name from default.test_data_04 limit 10) s1; --2 row update test_data_04 set name = 'adasd' where id = '1'; ERROR: 2024-05-23T10:26:32,028 ERROR [HiveServer2-Background-Pool: Thread-297] hive.HiveIcebergStorageHandler: Error while trying to commit job: job_17061635207991_169536, job_17061635207990_169536, job_17061635207992_169536, starting rollback changes for table: default.test_data_04 org.apache.iceberg.exceptions.NoSuchTableException: Table does not exist at location: /iceberg-catalog/warehouse/default/test_data_04 BEFORE UPDATE: ICEBERG TABLE DIR: [root@ ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 Found 2 items drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 /iceberg-catalog/warehouse/default/test_data_04/data drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 /iceberg-catalog/warehouse/default/test_data_04/metadata AFTER UPDATE: ICEBERG TABLE DIR: [root@XXX ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 Found 3 items drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_1 drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_2 drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_3 {code} was: Currently, when I update the location_based_table using hive, hive incorrectly empties all data directories and metadata directories. After the update statement is executed, the iceberg table is corrupted. {code:java} --spark: CREATE TABLE IF NOT EXISTS datacenter.default.test_data_04 ( id string,name string ) using iceberg PARTITIONED BY (name) TBLPROPERTIES ('read.orc.vectorization.enabled'='true','write.format.default'='orc','write.orc.bloom.filter.columns'='id','write.orc.compression-codec'='zstd','write.metadata.previous-versions-max'='3','write.metadata.delete-after-commit.enabled'='true'); insert into datacenter.default.test_data_04(id,name) values('1','a'),('2','b'); --hive4: CREATE EXTERNAL TABLE default.test_data_04 STORED BY 'org.apache.iceberg.mr.hive.HiveIcebergStorageHandler' LOCATION 'hdfs:///iceberg-catalog/warehouse/default/test_data_04' TBLPROPERTIES ('iceberg.catalog'='location_based_table','engine.hive.enabled'='true'); select distinct id,name from (select id,name from default.test_data_04 limit 10) s1; --2 row update test_data_04 set name = 'adasd' where id = '1'; ERROR: 2024-05-23T10:26:32,028 ERROR [HiveServer2-Background-Pool: Thread-297] hive.HiveIcebergStorageHandler: Error while trying to commit job: job_17061635207991_169536, job_17061635207990_169536, job_17061635207992_169536, starting rollback changes for table: default.test_data_04 org.apache.iceberg.exceptions.NoSuchTableException: Table does not exist at location: /iceberg-catalog/warehouse/default/test_data_04 BEFORE UPDATE: ICEBERG TABLE DIR: [root@ ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 Found 2 items drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 /iceberg-catalog/warehouse/default/test_data_04/data drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 /iceberg-catalog/warehouse/default/test_data_04/metadata AFTER UPDATE: ICEBERG TABLE DIR: [root@XXX ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 Found 3 items drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_1 drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_2 drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 /iceberg-catalog/warehouse
[jira] [Updated] (HIVE-28277) HIVE does not support update operations for ICEBERG of type location_based_table.
[ https://issues.apache.org/jira/browse/HIVE-28277?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] yongzhi.shao updated HIVE-28277: Description: Currently, when I update the location_based_table using hive, hive incorrectly empties all data directories and metadata directories. After the update statement is executed, the iceberg table is corrupted. {code:java} --spark: CREATE TABLE IF NOT EXISTS datacenter.default.test_data_04 ( id string,name string ) using iceberg PARTITIONED BY (name) TBLPROPERTIES ('read.orc.vectorization.enabled'='true','write.format.default'='orc','write.orc.bloom.filter.columns'='id','write.orc.compression-codec'='zstd','write.metadata.previous-versions-max'='3','write.metadata.delete-after-commit.enabled'='true'); insert into datacenter.default.test_data_04(id,name) values('1','a'),('2','b'); --hive4: CREATE EXTERNAL TABLE default.test_data_04 STORED BY 'org.apache.iceberg.mr.hive.HiveIcebergStorageHandler' LOCATION 'hdfs:///iceberg-catalog/warehouse/default/test_data_04' TBLPROPERTIES ('iceberg.catalog'='location_based_table','engine.hive.enabled'='true'); select distinct id,name from (select id,name from default.test_data_04 limit 10) s1; --2 row update test_data_04 set name = 'adasd' where id = '1'; ERROR: 2024-05-23T10:26:32,028 ERROR [HiveServer2-Background-Pool: Thread-297] hive.HiveIcebergStorageHandler: Error while trying to commit job: job_17061635207991_169536, job_17061635207990_169536, job_17061635207992_169536, starting rollback changes for table: default.test_data_04 org.apache.iceberg.exceptions.NoSuchTableException: Table does not exist at location: /iceberg-catalog/warehouse/default/test_data_04 BEFORE UPDATE: ICEBERG TABLE DIR: [root@ ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 Found 2 items drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 /iceberg-catalog/warehouse/default/test_data_04/data drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 /iceberg-catalog/warehouse/default/test_data_04/metadata AFTER UPDATE: ICEBERG TABLE DIR: [root@XXX ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 Found 3 items drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_1 drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_2 drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_3 {code} was: Currently, when I update the location_based_table using hive, hive incorrectly empties all data directories and metadata directories. {code:java} --spark: CREATE TABLE IF NOT EXISTS datacenter.default.test_data_04 ( id string,name string ) using iceberg PARTITIONED BY (name) TBLPROPERTIES ('read.orc.vectorization.enabled'='true','write.format.default'='orc','write.orc.bloom.filter.columns'='id','write.orc.compression-codec'='zstd','write.metadata.previous-versions-max'='3','write.metadata.delete-after-commit.enabled'='true'); insert into datacenter.default.test_data_04(id,name) values('1','a'),('2','b'); --hive4: CREATE EXTERNAL TABLE default.test_data_04 STORED BY 'org.apache.iceberg.mr.hive.HiveIcebergStorageHandler' LOCATION 'hdfs:///iceberg-catalog/warehouse/default/test_data_04' TBLPROPERTIES ('iceberg.catalog'='location_based_table','engine.hive.enabled'='true'); select distinct id,name from (select id,name from default.test_data_04 limit 10) s1; --2 row update test_data_04 set name = 'adasd' where id = '1'; ERROR: 2024-05-23T10:26:32,028 ERROR [HiveServer2-Background-Pool: Thread-297] hive.HiveIcebergStorageHandler: Error while trying to commit job: job_17061635207991_169536, job_17061635207990_169536, job_17061635207992_169536, starting rollback changes for table: default.test_data_04 org.apache.iceberg.exceptions.NoSuchTableException: Table does not exist at location: /iceberg-catalog/warehouse/default/test_data_04 BEFORE UPDATE: ICEBERG TABLE DIR: [root@ ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 Found 2 items drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 /iceberg-catalog/warehouse/default/test_data_04/data drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 /iceberg-catalog/warehouse/default/test_data_04/metadata AFTER UPDATE: ICEBERG TABLE DIR: [root@XXX ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 Found 3 items drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_1 drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_2 drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_3 {code} > HIVE does not support upd
[jira] [Created] (HIVE-28277) HIVE does not support update operations for ICEBERG of type location_based_table.
yongzhi.shao created HIVE-28277: --- Summary: HIVE does not support update operations for ICEBERG of type location_based_table. Key: HIVE-28277 URL: https://issues.apache.org/jira/browse/HIVE-28277 Project: Hive Issue Type: Improvement Components: Iceberg integration Affects Versions: 4.0.0 Environment: ICEBERG:1.5.2 HIVE 4.0.0 Reporter: yongzhi.shao Currently, when I update the location_based_table using hive, hive incorrectly empties all data directories and metadata directories. {code:java} --spark: CREATE TABLE IF NOT EXISTS datacenter.default.test_data_04 ( id string,name string ) using iceberg PARTITIONED BY (name) TBLPROPERTIES ('read.orc.vectorization.enabled'='true','write.format.default'='orc','write.orc.bloom.filter.columns'='id','write.orc.compression-codec'='zstd','write.metadata.previous-versions-max'='3','write.metadata.delete-after-commit.enabled'='true'); insert into datacenter.default.test_data_04(id,name) values('1','a'),('2','b'); --hive4: CREATE EXTERNAL TABLE default.test_data_04 STORED BY 'org.apache.iceberg.mr.hive.HiveIcebergStorageHandler' LOCATION 'hdfs:///iceberg-catalog/warehouse/default/test_data_04' TBLPROPERTIES ('iceberg.catalog'='location_based_table','engine.hive.enabled'='true'); select distinct id,name from (select id,name from default.test_data_04 limit 10) s1; --2 row update test_data_04 set name = 'adasd' where id = '1'; ERROR: 2024-05-23T10:26:32,028 ERROR [HiveServer2-Background-Pool: Thread-297] hive.HiveIcebergStorageHandler: Error while trying to commit job: job_17061635207991_169536, job_17061635207990_169536, job_17061635207992_169536, starting rollback changes for table: default.test_data_04 org.apache.iceberg.exceptions.NoSuchTableException: Table does not exist at location: /iceberg-catalog/warehouse/default/test_data_04 BEFORE UPDATE: ICEBERG TABLE DIR: [root@ ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 Found 2 items drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 /iceberg-catalog/warehouse/default/test_data_04/data drwxr-xr-x - hive hdfs 0 2024-05-23 09:26 /iceberg-catalog/warehouse/default/test_data_04/metadata AFTER UPDATE: ICEBERG TABLE DIR: [root@XXX ~]# hdfs dfs -ls /iceberg-catalog/warehouse/default/test_data_04 Found 3 items drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_1 drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_2 drwxr-xr-x - hive hdfs 0 2024-05-23 10:26 /iceberg-catalog/warehouse/default/test_data_04/-tmp.HIVE_UNION_SUBDIR_3 {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28276) Iceberg: Make Iceberg split threads configurable when table scanning
[ https://issues.apache.org/jira/browse/HIVE-28276?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated HIVE-28276: -- Labels: pull-request-available (was: ) > Iceberg: Make Iceberg split threads configurable when table scanning > > > Key: HIVE-28276 > URL: https://issues.apache.org/jira/browse/HIVE-28276 > Project: Hive > Issue Type: Improvement > Components: Iceberg integration >Reporter: Butao Zhang >Assignee: Butao Zhang >Priority: Major > Labels: pull-request-available > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28276) Iceberg: Make Iceberg split threads configurable when table scanning
Butao Zhang created HIVE-28276: -- Summary: Iceberg: Make Iceberg split threads configurable when table scanning Key: HIVE-28276 URL: https://issues.apache.org/jira/browse/HIVE-28276 Project: Hive Issue Type: Improvement Components: Iceberg integration Reporter: Butao Zhang -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (HIVE-28276) Iceberg: Make Iceberg split threads configurable when table scanning
[ https://issues.apache.org/jira/browse/HIVE-28276?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Butao Zhang reassigned HIVE-28276: -- Assignee: Butao Zhang > Iceberg: Make Iceberg split threads configurable when table scanning > > > Key: HIVE-28276 > URL: https://issues.apache.org/jira/browse/HIVE-28276 > Project: Hive > Issue Type: Improvement > Components: Iceberg integration >Reporter: Butao Zhang >Assignee: Butao Zhang >Priority: Major > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (HIVE-25351) stddev(), stddev_pop() with CBO enable returning null
[ https://issues.apache.org/jira/browse/HIVE-25351?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jiandan Yang reassigned HIVE-25351: Assignee: Jiandan Yang (was: Dayakar M) > stddev(), stddev_pop() with CBO enable returning null > - > > Key: HIVE-25351 > URL: https://issues.apache.org/jira/browse/HIVE-25351 > Project: Hive > Issue Type: Bug >Reporter: Ashish Sharma >Assignee: Jiandan Yang >Priority: Blocker > Labels: pull-request-available > > *script used to repro* > create table cbo_test (key string, v1 double, v2 decimal(30,2), v3 > decimal(30,2)); > insert into cbo_test values ("00140006375905", 10230.72, > 10230.72, 10230.69), ("00140006375905", 10230.72, 10230.72, > 10230.69), ("00140006375905", 10230.72, 10230.72, 10230.69), > ("00140006375905", 10230.72, 10230.72, 10230.69), > ("00140006375905", 10230.72, 10230.72, 10230.69), > ("00140006375905", 10230.72, 10230.72, 10230.69); > select stddev(v1), stddev(v2), stddev(v3) from cbo_test; > *Enable CBO* > ++ > | Explain | > ++ > | Plan optimized by CBO. | > || > | Vertex dependency in root stage| > | Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE)| > || > | Stage-0| > | Fetch Operator | > | limit:-1 | > | Stage-1| > | Reducer 2 vectorized | > | File Output Operator [FS_13] | > | Select Operator [SEL_12] (rows=1 width=24) | > | Output:["_col0","_col1","_col2"] | > | Group By Operator [GBY_11] (rows=1 width=72) | > | > Output:["_col0","_col1","_col2","_col3","_col4","_col5","_col6","_col7","_col8"],aggregations:["sum(VALUE._col0)","sum(VALUE._col1)","count(VALUE._col2)","sum(VALUE._col3)","sum(VALUE._col4)","count(VALUE._col5)","sum(VALUE._col6)","sum(VALUE._col7)","count(VALUE._col8)"] > | > | <-Map 1 [CUSTOM_SIMPLE_EDGE] vectorized | > | PARTITION_ONLY_SHUFFLE [RS_10] | > | Group By Operator [GBY_9] (rows=1 width=72) | > | > Output:["_col0","_col1","_col2","_col3","_col4","_col5","_col6","_col7","_col8"],aggregations:["sum(_col3)","sum(_col0)","count(_col0)","sum(_col5)","sum(_col4)","count(_col1)","sum(_col7)","sum(_col6)","count(_col2)"] > | > | Select Operator [SEL_8] (rows=6 width=232) | > | > Output:["_col0","_col1","_col2","_col3","_col4","_col5","_col6","_col7"] | > | TableScan [TS_0] (rows=6 width=232) | > | default@cbo_test,cbo_test, ACID > table,Tbl:COMPLETE,Col:COMPLETE,Output:["v1","v2","v3"] | > || > ++ > *Query Result* > _c0 _c1 _c2 > 0.0 NaN NaN > *Disable CBO* > ++ > | Explain | > ++ > | Vertex dependency in root stage| > | Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE)| > || > | Stage-0| > | Fetch Operator | > | limit:-1 | > | Stage-1| > | Reducer 2 vectorized | > | File Output Operator [FS_11]
[jira] [Updated] (HIVE-28274) Iceberg: Add support for 'If Not Exists' and 'or Replace' for Create Branch
[ https://issues.apache.org/jira/browse/HIVE-28274?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated HIVE-28274: -- Labels: pull-request-available (was: ) > Iceberg: Add support for 'If Not Exists' and 'or Replace' for Create Branch > --- > > Key: HIVE-28274 > URL: https://issues.apache.org/jira/browse/HIVE-28274 > Project: Hive > Issue Type: Sub-task >Reporter: Ayush Saxena >Assignee: Ayush Saxena >Priority: Major > Labels: pull-request-available > > Add support for > {noformat} > -- CREATE audit-branch at current snapshot with default retention if it > doesn't exist. > ALTER TABLE prod.db.sample CREATE BRANCH IF NOT EXISTS `audit-branch` > -- CREATE audit-branch at current snapshot with default retention or REPLACE > it if it already exists. > ALTER TABLE prod.db.sample CREATE OR REPLACE BRANCH `audit-branch`{noformat} > Like Spark: > https://iceberg.apache.org/docs/1.5.1/spark-ddl/#branching-and-tagging-ddl -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28274) Iceberg: Add support for 'If Not Exists' and 'or Replace' for Create Branch
[ https://issues.apache.org/jira/browse/HIVE-28274?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ayush Saxena updated HIVE-28274: Summary: Iceberg: Add support for 'If Not Exists' and 'or Replace' for Create Branch (was: Iceberg: Add support for 'If Not Exists" and 'or Replace' for Create Branch) > Iceberg: Add support for 'If Not Exists' and 'or Replace' for Create Branch > --- > > Key: HIVE-28274 > URL: https://issues.apache.org/jira/browse/HIVE-28274 > Project: Hive > Issue Type: Sub-task >Reporter: Ayush Saxena >Assignee: Ayush Saxena >Priority: Major > > Add support for > {noformat} > -- CREATE audit-branch at current snapshot with default retention if it > doesn't exist. > ALTER TABLE prod.db.sample CREATE BRANCH IF NOT EXISTS `audit-branch` > -- CREATE audit-branch at current snapshot with default retention or REPLACE > it if it already exists. > ALTER TABLE prod.db.sample CREATE OR REPLACE BRANCH `audit-branch`{noformat} > Like Spark: > https://iceberg.apache.org/docs/1.5.1/spark-ddl/#branching-and-tagging-ddl -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28275) Iceberg: Add support for 'If Not Exists" and 'or Replace' for Create Tag
Ayush Saxena created HIVE-28275: --- Summary: Iceberg: Add support for 'If Not Exists" and 'or Replace' for Create Tag Key: HIVE-28275 URL: https://issues.apache.org/jira/browse/HIVE-28275 Project: Hive Issue Type: Sub-task Reporter: Ayush Saxena Assignee: Ayush Saxena Add support for If not exists and Or Replace while creating Tags {noformat} -- CREATE historical-tag at current snapshot with default retention if it doesn't exist. ALTER TABLE prod.db.sample CREATE TAG IF NOT EXISTS `historical-tag` -- CREATE historical-tag at current snapshot with default retention or REPLACE it if it already exists. ALTER TABLE prod.db.sample CREATE OR REPLACE TAG `historical-tag`{noformat} Like Spark: https://iceberg.apache.org/docs/1.5.1/spark-ddl/#alter-table-create-branch -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28274) Iceberg: Add support for 'If Not Exists" and 'or Replace' for Create Branch
Ayush Saxena created HIVE-28274: --- Summary: Iceberg: Add support for 'If Not Exists" and 'or Replace' for Create Branch Key: HIVE-28274 URL: https://issues.apache.org/jira/browse/HIVE-28274 Project: Hive Issue Type: Sub-task Reporter: Ayush Saxena Assignee: Ayush Saxena Add support for {noformat} -- CREATE audit-branch at current snapshot with default retention if it doesn't exist. ALTER TABLE prod.db.sample CREATE BRANCH IF NOT EXISTS `audit-branch` -- CREATE audit-branch at current snapshot with default retention or REPLACE it if it already exists. ALTER TABLE prod.db.sample CREATE OR REPLACE BRANCH `audit-branch`{noformat} Like Spark: https://iceberg.apache.org/docs/1.5.1/spark-ddl/#branching-and-tagging-ddl -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Work started] (HIVE-28273) Test data generation failure in HIVE-28249 related tests
[ https://issues.apache.org/jira/browse/HIVE-28273?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Work on HIVE-28273 started by Csaba Juhász. --- > Test data generation failure in HIVE-28249 related tests > > > Key: HIVE-28273 > URL: https://issues.apache.org/jira/browse/HIVE-28273 > Project: Hive > Issue Type: Bug >Reporter: Csaba Juhász >Assignee: Csaba Juhász >Priority: Major > Labels: pull-request-available > Attachments: image-2024-05-22-19-11-35-890.png > > > generateJulianLeapYearTimestamps and generateJulianLeapYearTimestamps28thFeb > are throwing NegativeArraySizeException once the base value equals or is over > 999 > This is caused by the below code, supplying a negative value (when digits > return a value larger than 4) to zeros, which in turn is used to create a new > char array. > {code:java} > StringBuilder sb = new StringBuilder(29); > int year = ((i % ) + 1) * 100; > sb.append(zeros(4 - digits(year))); > {code} > When the tests are run using maven, the error in the generation function is > caught but never rethrown or reported and the build is reported successful. > For example running > _TestParquetTimestampsHive2Compatibility#testWriteHive2ReadHive4UsingLegacyConversionWithJulianLeapYearsFor28thFeb_ > has the result: > {code:java} > [INFO] --- > [INFO] T E S T S > [INFO] --- > [INFO] Running > org.apache.hadoop.hive.ql.io.parquet.serde.TestParquetTimestampsHive2Compatibility > [INFO] Tests run: 396, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: > 0.723 s - in > org.apache.hadoop.hive.ql.io.parquet.serde.TestParquetTimestampsHive2Compatibility > [INFO] > [INFO] Results: > [INFO] > [INFO] Tests run: 396, Failures: 0, Errors: 0, Skipped: 0 > ... > [INFO] BUILD SUCCESS > {code} > When the test is run through an IDE (eg VSCode), the failure is reported > properly. > !image-2024-05-22-19-11-35-890.png! -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (HIVE-28273) Test data generation failure in HIVE-28249 related tests
[ https://issues.apache.org/jira/browse/HIVE-28273?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Csaba Juhász reassigned HIVE-28273: --- Assignee: Csaba Juhász > Test data generation failure in HIVE-28249 related tests > > > Key: HIVE-28273 > URL: https://issues.apache.org/jira/browse/HIVE-28273 > Project: Hive > Issue Type: Bug >Reporter: Csaba Juhász >Assignee: Csaba Juhász >Priority: Major > Labels: pull-request-available > Attachments: image-2024-05-22-19-11-35-890.png > > > generateJulianLeapYearTimestamps and generateJulianLeapYearTimestamps28thFeb > are throwing NegativeArraySizeException once the base value equals or is over > 999 > This is caused by the below code, supplying a negative value (when digits > return a value larger than 4) to zeros, which in turn is used to create a new > char array. > {code:java} > StringBuilder sb = new StringBuilder(29); > int year = ((i % ) + 1) * 100; > sb.append(zeros(4 - digits(year))); > {code} > When the tests are run using maven, the error in the generation function is > caught but never rethrown or reported and the build is reported successful. > For example running > _TestParquetTimestampsHive2Compatibility#testWriteHive2ReadHive4UsingLegacyConversionWithJulianLeapYearsFor28thFeb_ > has the result: > {code:java} > [INFO] --- > [INFO] T E S T S > [INFO] --- > [INFO] Running > org.apache.hadoop.hive.ql.io.parquet.serde.TestParquetTimestampsHive2Compatibility > [INFO] Tests run: 396, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: > 0.723 s - in > org.apache.hadoop.hive.ql.io.parquet.serde.TestParquetTimestampsHive2Compatibility > [INFO] > [INFO] Results: > [INFO] > [INFO] Tests run: 396, Failures: 0, Errors: 0, Skipped: 0 > ... > [INFO] BUILD SUCCESS > {code} > When the test is run through an IDE (eg VSCode), the failure is reported > properly. > !image-2024-05-22-19-11-35-890.png! -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28273) Test data generation failure in HIVE-28249 related tests
[ https://issues.apache.org/jira/browse/HIVE-28273?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated HIVE-28273: -- Labels: pull-request-available (was: ) > Test data generation failure in HIVE-28249 related tests > > > Key: HIVE-28273 > URL: https://issues.apache.org/jira/browse/HIVE-28273 > Project: Hive > Issue Type: Bug >Reporter: Csaba Juhász >Priority: Major > Labels: pull-request-available > Attachments: image-2024-05-22-19-11-35-890.png > > > generateJulianLeapYearTimestamps and generateJulianLeapYearTimestamps28thFeb > are throwing NegativeArraySizeException once the base value equals or is over > 999 > This is caused by the below code, supplying a negative value (when digits > return a value larger than 4) to zeros, which in turn is used to create a new > char array. > {code:java} > StringBuilder sb = new StringBuilder(29); > int year = ((i % ) + 1) * 100; > sb.append(zeros(4 - digits(year))); > {code} > When the tests are run using maven, the error in the generation function is > caught but never rethrown or reported and the build is reported successful. > For example running > _TestParquetTimestampsHive2Compatibility#testWriteHive2ReadHive4UsingLegacyConversionWithJulianLeapYearsFor28thFeb_ > has the result: > {code:java} > [INFO] --- > [INFO] T E S T S > [INFO] --- > [INFO] Running > org.apache.hadoop.hive.ql.io.parquet.serde.TestParquetTimestampsHive2Compatibility > [INFO] Tests run: 396, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: > 0.723 s - in > org.apache.hadoop.hive.ql.io.parquet.serde.TestParquetTimestampsHive2Compatibility > [INFO] > [INFO] Results: > [INFO] > [INFO] Tests run: 396, Failures: 0, Errors: 0, Skipped: 0 > ... > [INFO] BUILD SUCCESS > {code} > When the test is run through an IDE (eg VSCode), the failure is reported > properly. > !image-2024-05-22-19-11-35-890.png! -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (HIVE-28270) Fix missing partition paths bug on drop_database
[ https://issues.apache.org/jira/browse/HIVE-28270?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17848691#comment-17848691 ] Ayush Saxena commented on HIVE-28270: - Committed to master. Thanx [~wechar] for the contribution!!! > Fix missing partition paths bug on drop_database > - > > Key: HIVE-28270 > URL: https://issues.apache.org/jira/browse/HIVE-28270 > Project: Hive > Issue Type: Bug > Components: Hive >Reporter: Wechar >Assignee: Wechar >Priority: Major > Labels: pull-request-available > > In {{HMSHandler#drop_database_core}}, it needs to collect all partition paths > that were not in the subdirectory of the table path, but now it only fetch > the last batch of paths. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28270) Fix missing partition paths bug on drop_database
[ https://issues.apache.org/jira/browse/HIVE-28270?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ayush Saxena updated HIVE-28270: Labels: hive-4.0.1-must pull-request-available (was: pull-request-available) > Fix missing partition paths bug on drop_database > - > > Key: HIVE-28270 > URL: https://issues.apache.org/jira/browse/HIVE-28270 > Project: Hive > Issue Type: Bug > Components: Hive >Reporter: Wechar >Assignee: Wechar >Priority: Major > Labels: hive-4.0.1-must, pull-request-available > Fix For: 4.1.0 > > > In {{HMSHandler#drop_database_core}}, it needs to collect all partition paths > that were not in the subdirectory of the table path, but now it only fetch > the last batch of paths. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (HIVE-28270) Fix missing partition paths bug on drop_database
[ https://issues.apache.org/jira/browse/HIVE-28270?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ayush Saxena resolved HIVE-28270. - Fix Version/s: 4.1.0 Resolution: Fixed > Fix missing partition paths bug on drop_database > - > > Key: HIVE-28270 > URL: https://issues.apache.org/jira/browse/HIVE-28270 > Project: Hive > Issue Type: Bug > Components: Hive >Reporter: Wechar >Assignee: Wechar >Priority: Major > Labels: pull-request-available > Fix For: 4.1.0 > > > In {{HMSHandler#drop_database_core}}, it needs to collect all partition paths > that were not in the subdirectory of the table path, but now it only fetch > the last batch of paths. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28271) DirectSql fails for AlterPartitions
[ https://issues.apache.org/jira/browse/HIVE-28271?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ayush Saxena updated HIVE-28271: Labels: hive-4.0.1-must pull-request-available (was: pull-request-available) > DirectSql fails for AlterPartitions > --- > > Key: HIVE-28271 > URL: https://issues.apache.org/jira/browse/HIVE-28271 > Project: Hive > Issue Type: Bug >Reporter: Ayush Saxena >Assignee: Ayush Saxena >Priority: Major > Labels: hive-4.0.1-must, pull-request-available > Fix For: 4.1.0 > > > It fails at three places: (Misses Database Which Uses CLOB & Missing Boolean > type conversions Checks > *First:* > {noformat} > 2024-05-21T08:50:16,570 WARN [main] metastore.ObjectStore: Falling back to > ORM path due to direct SQL failure (this is not an error): > java.lang.ClassCastException: org.apache.derby.impl.jdbc.EmbedClob cannot be > cast to java.lang.String at > org.apache.hadoop.hive.metastore.ExceptionHandler.newMetaException(ExceptionHandler.java:152) > at org.apache.hadoop.hive.metastore.Batchable.runBatched(Batchable.java:92) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.getParams(DirectSqlUpdatePart.java:748) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateParamTableInBatch(DirectSqlUpdatePart.java:715) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.alterPartitions(DirectSqlUpdatePart.java:636) > at > org.apache.hadoop.hive.metastore.MetaStoreDirectSql.alterPartitions(MetaStoreDirectSql.java:599) > at > org.apache.hadoop.hive.metastore.ObjectStore$20.getSqlResult(ObjectStore.java:5371); > {noformat} > *Second:* > {noformat} > 2024-05-21T09:14:36,808 WARN [main] metastore.ObjectStore: Falling back to > ORM path due to direct SQL failure (this is not an error): > java.lang.ClassCastException: org.apache.derby.impl.jdbc.EmbedClob cannot be > cast to java.lang.String at > org.apache.hadoop.hive.metastore.ExceptionHandler.newMetaException(ExceptionHandler.java:152) > at org.apache.hadoop.hive.metastore.Batchable.runBatched(Batchable.java:92) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateCDInBatch(DirectSqlUpdatePart.java:1228) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateStorageDescriptorInBatch(DirectSqlUpdatePart.java:888) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.alterPartitions(DirectSqlUpdatePart.java:638) > at > org.apache.hadoop.hive.metastore.MetaStoreDirectSql.alterPartitions(MetaStoreDirectSql.java:599) > at > org.apache.hadoop.hive.metastore.ObjectStore$20.getSqlResult(ObjectStore.java:5371);{noformat} > *Third: Missing Boolean check type* > {noformat} > 2024-05-21T09:35:44,063 WARN [main] metastore.ObjectStore: Falling back to > ORM path due to direct SQL failure (this is not an error): > java.sql.BatchUpdateException: A truncation error was encountered trying to > shrink CHAR 'false' to length 1. at > org.apache.hadoop.hive.metastore.ExceptionHandler.newMetaException(ExceptionHandler.java:152) > at org.apache.hadoop.hive.metastore.Batchable.runBatched(Batchable.java:92) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.lambda$updateSDInBatch$16(DirectSqlUpdatePart.java:926) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateWithStatement(DirectSqlUpdatePart.java:656) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateSDInBatch(DirectSqlUpdatePart.java:926) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateStorageDescriptorInBatch(DirectSqlUpdatePart.java:900) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.alterPartitions(DirectSqlUpdatePart.java:638) > at > org.apache.hadoop.hive.metastore.MetaStoreDirectSql.alterPartitions(MetaStoreDirectSql.java:599) > at > org.apache.hadoop.hive.metastore.ObjectStore$20.getSqlResult(ObjectStore.java:5371); > {noformat} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (HIVE-28271) DirectSql fails for AlterPartitions
[ https://issues.apache.org/jira/browse/HIVE-28271?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ayush Saxena resolved HIVE-28271. - Fix Version/s: 4.1.0 Resolution: Fixed > DirectSql fails for AlterPartitions > --- > > Key: HIVE-28271 > URL: https://issues.apache.org/jira/browse/HIVE-28271 > Project: Hive > Issue Type: Bug >Reporter: Ayush Saxena >Assignee: Ayush Saxena >Priority: Major > Labels: pull-request-available > Fix For: 4.1.0 > > > It fails at three places: (Misses Database Which Uses CLOB & Missing Boolean > type conversions Checks > *First:* > {noformat} > 2024-05-21T08:50:16,570 WARN [main] metastore.ObjectStore: Falling back to > ORM path due to direct SQL failure (this is not an error): > java.lang.ClassCastException: org.apache.derby.impl.jdbc.EmbedClob cannot be > cast to java.lang.String at > org.apache.hadoop.hive.metastore.ExceptionHandler.newMetaException(ExceptionHandler.java:152) > at org.apache.hadoop.hive.metastore.Batchable.runBatched(Batchable.java:92) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.getParams(DirectSqlUpdatePart.java:748) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateParamTableInBatch(DirectSqlUpdatePart.java:715) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.alterPartitions(DirectSqlUpdatePart.java:636) > at > org.apache.hadoop.hive.metastore.MetaStoreDirectSql.alterPartitions(MetaStoreDirectSql.java:599) > at > org.apache.hadoop.hive.metastore.ObjectStore$20.getSqlResult(ObjectStore.java:5371); > {noformat} > *Second:* > {noformat} > 2024-05-21T09:14:36,808 WARN [main] metastore.ObjectStore: Falling back to > ORM path due to direct SQL failure (this is not an error): > java.lang.ClassCastException: org.apache.derby.impl.jdbc.EmbedClob cannot be > cast to java.lang.String at > org.apache.hadoop.hive.metastore.ExceptionHandler.newMetaException(ExceptionHandler.java:152) > at org.apache.hadoop.hive.metastore.Batchable.runBatched(Batchable.java:92) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateCDInBatch(DirectSqlUpdatePart.java:1228) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateStorageDescriptorInBatch(DirectSqlUpdatePart.java:888) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.alterPartitions(DirectSqlUpdatePart.java:638) > at > org.apache.hadoop.hive.metastore.MetaStoreDirectSql.alterPartitions(MetaStoreDirectSql.java:599) > at > org.apache.hadoop.hive.metastore.ObjectStore$20.getSqlResult(ObjectStore.java:5371);{noformat} > *Third: Missing Boolean check type* > {noformat} > 2024-05-21T09:35:44,063 WARN [main] metastore.ObjectStore: Falling back to > ORM path due to direct SQL failure (this is not an error): > java.sql.BatchUpdateException: A truncation error was encountered trying to > shrink CHAR 'false' to length 1. at > org.apache.hadoop.hive.metastore.ExceptionHandler.newMetaException(ExceptionHandler.java:152) > at org.apache.hadoop.hive.metastore.Batchable.runBatched(Batchable.java:92) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.lambda$updateSDInBatch$16(DirectSqlUpdatePart.java:926) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateWithStatement(DirectSqlUpdatePart.java:656) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateSDInBatch(DirectSqlUpdatePart.java:926) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateStorageDescriptorInBatch(DirectSqlUpdatePart.java:900) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.alterPartitions(DirectSqlUpdatePart.java:638) > at > org.apache.hadoop.hive.metastore.MetaStoreDirectSql.alterPartitions(MetaStoreDirectSql.java:599) > at > org.apache.hadoop.hive.metastore.ObjectStore$20.getSqlResult(ObjectStore.java:5371); > {noformat} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (HIVE-28271) DirectSql fails for AlterPartitions
[ https://issues.apache.org/jira/browse/HIVE-28271?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17848690#comment-17848690 ] Ayush Saxena commented on HIVE-28271: - Committed to master. Thanx [~zhangbutao] & [~wechar] for the review!! > DirectSql fails for AlterPartitions > --- > > Key: HIVE-28271 > URL: https://issues.apache.org/jira/browse/HIVE-28271 > Project: Hive > Issue Type: Bug >Reporter: Ayush Saxena >Assignee: Ayush Saxena >Priority: Major > Labels: pull-request-available > > It fails at three places: (Misses Database Which Uses CLOB & Missing Boolean > type conversions Checks > *First:* > {noformat} > 2024-05-21T08:50:16,570 WARN [main] metastore.ObjectStore: Falling back to > ORM path due to direct SQL failure (this is not an error): > java.lang.ClassCastException: org.apache.derby.impl.jdbc.EmbedClob cannot be > cast to java.lang.String at > org.apache.hadoop.hive.metastore.ExceptionHandler.newMetaException(ExceptionHandler.java:152) > at org.apache.hadoop.hive.metastore.Batchable.runBatched(Batchable.java:92) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.getParams(DirectSqlUpdatePart.java:748) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateParamTableInBatch(DirectSqlUpdatePart.java:715) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.alterPartitions(DirectSqlUpdatePart.java:636) > at > org.apache.hadoop.hive.metastore.MetaStoreDirectSql.alterPartitions(MetaStoreDirectSql.java:599) > at > org.apache.hadoop.hive.metastore.ObjectStore$20.getSqlResult(ObjectStore.java:5371); > {noformat} > *Second:* > {noformat} > 2024-05-21T09:14:36,808 WARN [main] metastore.ObjectStore: Falling back to > ORM path due to direct SQL failure (this is not an error): > java.lang.ClassCastException: org.apache.derby.impl.jdbc.EmbedClob cannot be > cast to java.lang.String at > org.apache.hadoop.hive.metastore.ExceptionHandler.newMetaException(ExceptionHandler.java:152) > at org.apache.hadoop.hive.metastore.Batchable.runBatched(Batchable.java:92) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateCDInBatch(DirectSqlUpdatePart.java:1228) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateStorageDescriptorInBatch(DirectSqlUpdatePart.java:888) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.alterPartitions(DirectSqlUpdatePart.java:638) > at > org.apache.hadoop.hive.metastore.MetaStoreDirectSql.alterPartitions(MetaStoreDirectSql.java:599) > at > org.apache.hadoop.hive.metastore.ObjectStore$20.getSqlResult(ObjectStore.java:5371);{noformat} > *Third: Missing Boolean check type* > {noformat} > 2024-05-21T09:35:44,063 WARN [main] metastore.ObjectStore: Falling back to > ORM path due to direct SQL failure (this is not an error): > java.sql.BatchUpdateException: A truncation error was encountered trying to > shrink CHAR 'false' to length 1. at > org.apache.hadoop.hive.metastore.ExceptionHandler.newMetaException(ExceptionHandler.java:152) > at org.apache.hadoop.hive.metastore.Batchable.runBatched(Batchable.java:92) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.lambda$updateSDInBatch$16(DirectSqlUpdatePart.java:926) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateWithStatement(DirectSqlUpdatePart.java:656) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateSDInBatch(DirectSqlUpdatePart.java:926) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateStorageDescriptorInBatch(DirectSqlUpdatePart.java:900) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.alterPartitions(DirectSqlUpdatePart.java:638) > at > org.apache.hadoop.hive.metastore.MetaStoreDirectSql.alterPartitions(MetaStoreDirectSql.java:599) > at > org.apache.hadoop.hive.metastore.ObjectStore$20.getSqlResult(ObjectStore.java:5371); > {noformat} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28273) Test data generation failure in HIVE-28249 related tests
Csaba Juhász created HIVE-28273: --- Summary: Test data generation failure in HIVE-28249 related tests Key: HIVE-28273 URL: https://issues.apache.org/jira/browse/HIVE-28273 Project: Hive Issue Type: Bug Reporter: Csaba Juhász Attachments: image-2024-05-22-19-11-35-890.png generateJulianLeapYearTimestamps and generateJulianLeapYearTimestamps28thFeb are throwing NegativeArraySizeException once the base value equals or is over 999 This is caused by the below code, supplying a negative value (when digits return a value larger than 4) to zeros, which in turn is used to create a new char array. {code:java} StringBuilder sb = new StringBuilder(29); int year = ((i % ) + 1) * 100; sb.append(zeros(4 - digits(year))); {code} When the tests are run using maven, the error in the generation function is caught but never rethrown or reported and the build is reported successful. For example running _TestParquetTimestampsHive2Compatibility#testWriteHive2ReadHive4UsingLegacyConversionWithJulianLeapYearsFor28thFeb_ has the result: {code:java} [INFO] --- [INFO] T E S T S [INFO] --- [INFO] Running org.apache.hadoop.hive.ql.io.parquet.serde.TestParquetTimestampsHive2Compatibility [INFO] Tests run: 396, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.723 s - in org.apache.hadoop.hive.ql.io.parquet.serde.TestParquetTimestampsHive2Compatibility [INFO] [INFO] Results: [INFO] [INFO] Tests run: 396, Failures: 0, Errors: 0, Skipped: 0 ... [INFO] BUILD SUCCESS {code} When the test is run through an IDE (eg VSCode), the failure is reported properly. !image-2024-05-22-19-11-35-890.png! -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (HIVE-28246) Fix confusing log message in LlapTaskSchedulerService
[ https://issues.apache.org/jira/browse/HIVE-28246?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ayush Saxena resolved HIVE-28246. - Fix Version/s: 4.1.0 Resolution: Fixed > Fix confusing log message in LlapTaskSchedulerService > - > > Key: HIVE-28246 > URL: https://issues.apache.org/jira/browse/HIVE-28246 > Project: Hive > Issue Type: Improvement >Reporter: László Bodor >Assignee: Zoltán Rátkai >Priority: Major > Labels: newbie, pull-request-available > Fix For: 4.1.0 > > > https://github.com/apache/hive/blob/8415527101432bb5bf14b3c2a318a2cc40801b9a/llap-tez/src/java/org/apache/hadoop/hive/llap/tezplugins/LlapTaskSchedulerService.java#L1719 > {code} > WM_LOG.info("Registering " + taskInfo.attemptId + "; " + > taskInfo.isGuaranteed); > {code} > leads to a message like: > {code} > Registering attempt_1714730410273_0009_153_05_000235_10; false > {code} > "false" is out of any context, supposed to be something like: > {code} > Registering attempt_1714730410273_0009_153_05_000235_10, guaranteed: false > {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (HIVE-28246) Fix confusing log message in LlapTaskSchedulerService
[ https://issues.apache.org/jira/browse/HIVE-28246?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17848645#comment-17848645 ] Ayush Saxena commented on HIVE-28246: - Committed to master. Thanx [~zratkai] for the contribution & [~aturoczy] for the review!!! > Fix confusing log message in LlapTaskSchedulerService > - > > Key: HIVE-28246 > URL: https://issues.apache.org/jira/browse/HIVE-28246 > Project: Hive > Issue Type: Improvement >Reporter: László Bodor >Assignee: Zoltán Rátkai >Priority: Major > Labels: newbie, pull-request-available > > https://github.com/apache/hive/blob/8415527101432bb5bf14b3c2a318a2cc40801b9a/llap-tez/src/java/org/apache/hadoop/hive/llap/tezplugins/LlapTaskSchedulerService.java#L1719 > {code} > WM_LOG.info("Registering " + taskInfo.attemptId + "; " + > taskInfo.isGuaranteed); > {code} > leads to a message like: > {code} > Registering attempt_1714730410273_0009_153_05_000235_10; false > {code} > "false" is out of any context, supposed to be something like: > {code} > Registering attempt_1714730410273_0009_153_05_000235_10, guaranteed: false > {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28246) Fix confusing log message in LlapTaskSchedulerService
[ https://issues.apache.org/jira/browse/HIVE-28246?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ayush Saxena updated HIVE-28246: Summary: Fix confusing log message in LlapTaskSchedulerService (was: Confusing log messages in LlapTaskScheduler) > Fix confusing log message in LlapTaskSchedulerService > - > > Key: HIVE-28246 > URL: https://issues.apache.org/jira/browse/HIVE-28246 > Project: Hive > Issue Type: Improvement >Reporter: László Bodor >Assignee: Zoltán Rátkai >Priority: Major > Labels: newbie, pull-request-available > > https://github.com/apache/hive/blob/8415527101432bb5bf14b3c2a318a2cc40801b9a/llap-tez/src/java/org/apache/hadoop/hive/llap/tezplugins/LlapTaskSchedulerService.java#L1719 > {code} > WM_LOG.info("Registering " + taskInfo.attemptId + "; " + > taskInfo.isGuaranteed); > {code} > leads to a message like: > {code} > Registering attempt_1714730410273_0009_153_05_000235_10; false > {code} > "false" is out of any context, supposed to be something like: > {code} > Registering attempt_1714730410273_0009_153_05_000235_10, guaranteed: false > {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (HIVE-25974) Drop HiveFilterMergeRule and use FilterMergeRule from Calcite
[ https://issues.apache.org/jira/browse/HIVE-25974?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Stamatis Zampetakis resolved HIVE-25974. Fix Version/s: Not Applicable Resolution: Duplicate > Drop HiveFilterMergeRule and use FilterMergeRule from Calcite > - > > Key: HIVE-25974 > URL: https://issues.apache.org/jira/browse/HIVE-25974 > Project: Hive > Issue Type: Improvement > Components: CBO >Affects Versions: 4.0.0 >Reporter: Alessandro Solimando >Priority: Major > Fix For: Not Applicable > > > HiveFilterMergeRule is a copy of FilterMergeRule which was needed since the > latter did not simplify/flatten before creating the merged filter. > This behaviour has been fixed in CALCITE-3982 (released since 1.23), so it > seems that the Hive rule could be removed now. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Comment Edited] (HIVE-22633) GroupByOperator may throw NullPointerException when setting data skew optimization parameters
[ https://issues.apache.org/jira/browse/HIVE-22633?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17848545#comment-17848545 ] Butao Zhang edited comment on HIVE-22633 at 5/22/24 10:46 AM: -- Update: if you are using Hive3, you can try to patch this HIVE-27712 as a hotfix, which is easier to test than HIVE-23530. Hive4 does not have this issue as we don't use this udaf since HIVE-23530 . was (Author: zhangbutao): Update: if you are using Hive3, you can try to patch this HIVE-27712 as a hotfix, which is easier to test than HIVE-23530. > GroupByOperator may throw NullPointerException when setting data skew > optimization parameters > - > > Key: HIVE-22633 > URL: https://issues.apache.org/jira/browse/HIVE-22633 > Project: Hive > Issue Type: Bug >Affects Versions: 3.1.0, 3.1.1, 4.0.0 >Reporter: Butao Zhang >Assignee: Butao Zhang >Priority: Major > > if hive.map.aggr and hive.groupby.skewindata set true,exception will be > thrown. > step to repro: > 1. create table: > set hive.map.aggr=true; > set hive.groupby.skewindata=true; > create table test1 (id1 bigint); > create table test2 (id2 bigint) partitioned by(dt2 string); > insert into test2 partition(dt2='2020') select a.id1 from test1 a group by > a.id1; > 2.NullPointerException: > {code:java} > ], TaskAttempt 2 failed, info=[Error: Error while running task ( failure ) : > attempt_1585641455670_0001_2_03_00_2:java.lang.RuntimeException: > java.lang.NullPointerException > at > org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:296) > at > org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:250) > at > org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:374) > at > org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:73) > at > org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:61) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1682) > at > org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:61) > at > org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:37) > at org.apache.tez.common.CallableWithNdc.call(CallableWithNdc.java:36) > at > com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:108) > at > com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:41) > at > com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:77) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > at java.lang.Thread.run(Thread.java:748) > Caused by: java.lang.NullPointerException > at > org.apache.hadoop.hive.ql.udf.generic.GenericUDAFComputeStats$GenericUDAFNumericStatsEvaluator.init(GenericUDAFComputeStats.java:373) > at > org.apache.hadoop.hive.ql.exec.GroupByOperator.initializeOp(GroupByOperator.java:373) > at > org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:360) > at > org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.init(ReduceRecordProcessor.java:191) > at > org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:266) > {code} > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (HIVE-22633) GroupByOperator may throw NullPointerException when setting data skew optimization parameters
[ https://issues.apache.org/jira/browse/HIVE-22633?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17848545#comment-17848545 ] Butao Zhang commented on HIVE-22633: Update: if you are using Hive3, you can try to patch this HIVE-27712 as a hotfix, which is easier to test than HIVE-23530. > GroupByOperator may throw NullPointerException when setting data skew > optimization parameters > - > > Key: HIVE-22633 > URL: https://issues.apache.org/jira/browse/HIVE-22633 > Project: Hive > Issue Type: Bug >Affects Versions: 3.1.0, 3.1.1, 4.0.0 >Reporter: Butao Zhang >Assignee: Butao Zhang >Priority: Major > > if hive.map.aggr and hive.groupby.skewindata set true,exception will be > thrown. > step to repro: > 1. create table: > set hive.map.aggr=true; > set hive.groupby.skewindata=true; > create table test1 (id1 bigint); > create table test2 (id2 bigint) partitioned by(dt2 string); > insert into test2 partition(dt2='2020') select a.id1 from test1 a group by > a.id1; > 2.NullPointerException: > {code:java} > ], TaskAttempt 2 failed, info=[Error: Error while running task ( failure ) : > attempt_1585641455670_0001_2_03_00_2:java.lang.RuntimeException: > java.lang.NullPointerException > at > org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:296) > at > org.apache.hadoop.hive.ql.exec.tez.TezProcessor.run(TezProcessor.java:250) > at > org.apache.tez.runtime.LogicalIOProcessorRuntimeTask.run(LogicalIOProcessorRuntimeTask.java:374) > at > org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:73) > at > org.apache.tez.runtime.task.TaskRunner2Callable$1.run(TaskRunner2Callable.java:61) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1682) > at > org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:61) > at > org.apache.tez.runtime.task.TaskRunner2Callable.callInternal(TaskRunner2Callable.java:37) > at org.apache.tez.common.CallableWithNdc.call(CallableWithNdc.java:36) > at > com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:108) > at > com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:41) > at > com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:77) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > at java.lang.Thread.run(Thread.java:748) > Caused by: java.lang.NullPointerException > at > org.apache.hadoop.hive.ql.udf.generic.GenericUDAFComputeStats$GenericUDAFNumericStatsEvaluator.init(GenericUDAFComputeStats.java:373) > at > org.apache.hadoop.hive.ql.exec.GroupByOperator.initializeOp(GroupByOperator.java:373) > at > org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:360) > at > org.apache.hadoop.hive.ql.exec.tez.ReduceRecordProcessor.init(ReduceRecordProcessor.java:191) > at > org.apache.hadoop.hive.ql.exec.tez.TezProcessor.initializeAndRunProcessor(TezProcessor.java:266) > {code} > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28272) Support setting per-session S3 credentials in Warehouse
[ https://issues.apache.org/jira/browse/HIVE-28272?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated HIVE-28272: -- Labels: pull-request-available (was: ) > Support setting per-session S3 credentials in Warehouse > --- > > Key: HIVE-28272 > URL: https://issues.apache.org/jira/browse/HIVE-28272 > Project: Hive > Issue Type: Improvement >Reporter: Butao Zhang >Assignee: Butao Zhang >Priority: Major > Labels: pull-request-available > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (HIVE-28272) Support setting per-session S3 credentials in Warehouse
[ https://issues.apache.org/jira/browse/HIVE-28272?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Butao Zhang reassigned HIVE-28272: -- Assignee: Butao Zhang > Support setting per-session S3 credentials in Warehouse > --- > > Key: HIVE-28272 > URL: https://issues.apache.org/jira/browse/HIVE-28272 > Project: Hive > Issue Type: Improvement >Reporter: Butao Zhang >Assignee: Butao Zhang >Priority: Major > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (HIVE-28268) Iceberg: Retrieve row count from iceberg SnapshotSummary in case of iceberg.hive.keep.stats=false
[ https://issues.apache.org/jira/browse/HIVE-28268?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Butao Zhang reassigned HIVE-28268: -- Assignee: Butao Zhang > Iceberg: Retrieve row count from iceberg SnapshotSummary in case of > iceberg.hive.keep.stats=false > - > > Key: HIVE-28268 > URL: https://issues.apache.org/jira/browse/HIVE-28268 > Project: Hive > Issue Type: Task > Components: Iceberg integration >Reporter: Butao Zhang >Assignee: Butao Zhang >Priority: Major > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28272) Support setting per-session S3 credentials in Warehouse
Butao Zhang created HIVE-28272: -- Summary: Support setting per-session S3 credentials in Warehouse Key: HIVE-28272 URL: https://issues.apache.org/jira/browse/HIVE-28272 Project: Hive Issue Type: Improvement Reporter: Butao Zhang -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (HIVE-25351) stddev(), stddev_pop() with CBO enable returning null
[ https://issues.apache.org/jira/browse/HIVE-25351?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17848524#comment-17848524 ] Dayakar M commented on HIVE-25351: -- [~yangjiandan] currently I am not working on this issue, if you have a solution ready then you can take it over and fix it. Thanks. > stddev(), stddev_pop() with CBO enable returning null > - > > Key: HIVE-25351 > URL: https://issues.apache.org/jira/browse/HIVE-25351 > Project: Hive > Issue Type: Bug >Reporter: Ashish Sharma >Assignee: Dayakar M >Priority: Blocker > Labels: pull-request-available > > *script used to repro* > create table cbo_test (key string, v1 double, v2 decimal(30,2), v3 > decimal(30,2)); > insert into cbo_test values ("00140006375905", 10230.72, > 10230.72, 10230.69), ("00140006375905", 10230.72, 10230.72, > 10230.69), ("00140006375905", 10230.72, 10230.72, 10230.69), > ("00140006375905", 10230.72, 10230.72, 10230.69), > ("00140006375905", 10230.72, 10230.72, 10230.69), > ("00140006375905", 10230.72, 10230.72, 10230.69); > select stddev(v1), stddev(v2), stddev(v3) from cbo_test; > *Enable CBO* > ++ > | Explain | > ++ > | Plan optimized by CBO. | > || > | Vertex dependency in root stage| > | Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE)| > || > | Stage-0| > | Fetch Operator | > | limit:-1 | > | Stage-1| > | Reducer 2 vectorized | > | File Output Operator [FS_13] | > | Select Operator [SEL_12] (rows=1 width=24) | > | Output:["_col0","_col1","_col2"] | > | Group By Operator [GBY_11] (rows=1 width=72) | > | > Output:["_col0","_col1","_col2","_col3","_col4","_col5","_col6","_col7","_col8"],aggregations:["sum(VALUE._col0)","sum(VALUE._col1)","count(VALUE._col2)","sum(VALUE._col3)","sum(VALUE._col4)","count(VALUE._col5)","sum(VALUE._col6)","sum(VALUE._col7)","count(VALUE._col8)"] > | > | <-Map 1 [CUSTOM_SIMPLE_EDGE] vectorized | > | PARTITION_ONLY_SHUFFLE [RS_10] | > | Group By Operator [GBY_9] (rows=1 width=72) | > | > Output:["_col0","_col1","_col2","_col3","_col4","_col5","_col6","_col7","_col8"],aggregations:["sum(_col3)","sum(_col0)","count(_col0)","sum(_col5)","sum(_col4)","count(_col1)","sum(_col7)","sum(_col6)","count(_col2)"] > | > | Select Operator [SEL_8] (rows=6 width=232) | > | > Output:["_col0","_col1","_col2","_col3","_col4","_col5","_col6","_col7"] | > | TableScan [TS_0] (rows=6 width=232) | > | default@cbo_test,cbo_test, ACID > table,Tbl:COMPLETE,Col:COMPLETE,Output:["v1","v2","v3"] | > || > ++ > *Query Result* > _c0 _c1 _c2 > 0.0 NaN NaN > *Disable CBO* > ++ > | Explain | > ++ > | Vertex dependency in root stage| > | Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE)| > || > | Stage-0| > | Fetch Operator | > | limit:-1 | > | Stage-1
[jira] [Resolved] (HIVE-28266) Iceberg: select count(*) from data_files metadata tables gives wrong result
[ https://issues.apache.org/jira/browse/HIVE-28266?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Denys Kuzmenko resolved HIVE-28266. --- Fix Version/s: 4.1.0 Resolution: Fixed > Iceberg: select count(*) from data_files metadata tables gives wrong result > --- > > Key: HIVE-28266 > URL: https://issues.apache.org/jira/browse/HIVE-28266 > Project: Hive > Issue Type: Bug >Reporter: Dmitriy Fingerman >Assignee: Dmitriy Fingerman >Priority: Major > Labels: pull-request-available > Fix For: 4.1.0 > > > In Hive Iceberg, every table has a corresponding metadata table > "*.data_files" that contains info about the files that contain table's data. > select count(*) from a data_file metadata table returns number of rows in the > data table instead of number of data files from the metadata table. > > {code:java} > CREATE TABLE x (name VARCHAR(50), age TINYINT, num_clicks BIGINT) stored by > iceberg stored as orc TBLPROPERTIES > ('external.table.purge'='true','format-version'='2'); > insert into x values > ('amy', 35, 123412344), > ('adxfvy', 36, 123412534), > ('amsdfyy', 37, 123417234), > ('asafmy', 38, 123412534); > insert into x values > ('amerqwy', 39, 123441234), > ('amyxzcv', 40, 123341234), > ('erweramy', 45, 122341234); > Select * from default.x.data_files; > – Returns 2 records in the output > Select count from default.x.data_files; > – Returns 7 instead of 2 > {code} > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28266) Iceberg: select count(*) from data_files metadata tables gives wrong result
[ https://issues.apache.org/jira/browse/HIVE-28266?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Denys Kuzmenko updated HIVE-28266: -- Affects Version/s: 4.0.0 > Iceberg: select count(*) from data_files metadata tables gives wrong result > --- > > Key: HIVE-28266 > URL: https://issues.apache.org/jira/browse/HIVE-28266 > Project: Hive > Issue Type: Bug >Affects Versions: 4.0.0 >Reporter: Dmitriy Fingerman >Assignee: Dmitriy Fingerman >Priority: Major > Labels: pull-request-available > Fix For: 4.1.0 > > > In Hive Iceberg, every table has a corresponding metadata table > "*.data_files" that contains info about the files that contain table's data. > select count(*) from a data_file metadata table returns number of rows in the > data table instead of number of data files from the metadata table. > > {code:java} > CREATE TABLE x (name VARCHAR(50), age TINYINT, num_clicks BIGINT) stored by > iceberg stored as orc TBLPROPERTIES > ('external.table.purge'='true','format-version'='2'); > insert into x values > ('amy', 35, 123412344), > ('adxfvy', 36, 123412534), > ('amsdfyy', 37, 123417234), > ('asafmy', 38, 123412534); > insert into x values > ('amerqwy', 39, 123441234), > ('amyxzcv', 40, 123341234), > ('erweramy', 45, 122341234); > Select * from default.x.data_files; > – Returns 2 records in the output > Select count from default.x.data_files; > – Returns 7 instead of 2 > {code} > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (HIVE-28266) Iceberg: select count(*) from data_files metadata tables gives wrong result
[ https://issues.apache.org/jira/browse/HIVE-28266?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17848501#comment-17848501 ] Denys Kuzmenko commented on HIVE-28266: --- Merged to master Thanks [~difin] for the patch and [~zhangbutao] for the review! > Iceberg: select count(*) from data_files metadata tables gives wrong result > --- > > Key: HIVE-28266 > URL: https://issues.apache.org/jira/browse/HIVE-28266 > Project: Hive > Issue Type: Bug >Reporter: Dmitriy Fingerman >Assignee: Dmitriy Fingerman >Priority: Major > Labels: pull-request-available > > In Hive Iceberg, every table has a corresponding metadata table > "*.data_files" that contains info about the files that contain table's data. > select count(*) from a data_file metadata table returns number of rows in the > data table instead of number of data files from the metadata table. > > {code:java} > CREATE TABLE x (name VARCHAR(50), age TINYINT, num_clicks BIGINT) stored by > iceberg stored as orc TBLPROPERTIES > ('external.table.purge'='true','format-version'='2'); > insert into x values > ('amy', 35, 123412344), > ('adxfvy', 36, 123412534), > ('amsdfyy', 37, 123417234), > ('asafmy', 38, 123412534); > insert into x values > ('amerqwy', 39, 123441234), > ('amyxzcv', 40, 123341234), > ('erweramy', 45, 122341234); > Select * from default.x.data_files; > – Returns 2 records in the output > Select count from default.x.data_files; > – Returns 7 instead of 2 > {code} > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (HIVE-25351) stddev(), stddev_pop() with CBO enable returning null
[ https://issues.apache.org/jira/browse/HIVE-25351?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17848487#comment-17848487 ] Jiandan Yang commented on HIVE-25351: -- [~Dayakar] I encountered the same issue in Hive version 3.1.3, and from reviewing the code, it appears that the current master branch would have the same issue. I have fixed this problem in version 3.1.3. If no one is addressing this issue, I am prepared to take it over and resolve it. > stddev(), stddev_pop() with CBO enable returning null > - > > Key: HIVE-25351 > URL: https://issues.apache.org/jira/browse/HIVE-25351 > Project: Hive > Issue Type: Bug >Reporter: Ashish Sharma >Assignee: Dayakar M >Priority: Blocker > Labels: pull-request-available > > *script used to repro* > create table cbo_test (key string, v1 double, v2 decimal(30,2), v3 > decimal(30,2)); > insert into cbo_test values ("00140006375905", 10230.72, > 10230.72, 10230.69), ("00140006375905", 10230.72, 10230.72, > 10230.69), ("00140006375905", 10230.72, 10230.72, 10230.69), > ("00140006375905", 10230.72, 10230.72, 10230.69), > ("00140006375905", 10230.72, 10230.72, 10230.69), > ("00140006375905", 10230.72, 10230.72, 10230.69); > select stddev(v1), stddev(v2), stddev(v3) from cbo_test; > *Enable CBO* > ++ > | Explain | > ++ > | Plan optimized by CBO. | > || > | Vertex dependency in root stage| > | Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE)| > || > | Stage-0| > | Fetch Operator | > | limit:-1 | > | Stage-1| > | Reducer 2 vectorized | > | File Output Operator [FS_13] | > | Select Operator [SEL_12] (rows=1 width=24) | > | Output:["_col0","_col1","_col2"] | > | Group By Operator [GBY_11] (rows=1 width=72) | > | > Output:["_col0","_col1","_col2","_col3","_col4","_col5","_col6","_col7","_col8"],aggregations:["sum(VALUE._col0)","sum(VALUE._col1)","count(VALUE._col2)","sum(VALUE._col3)","sum(VALUE._col4)","count(VALUE._col5)","sum(VALUE._col6)","sum(VALUE._col7)","count(VALUE._col8)"] > | > | <-Map 1 [CUSTOM_SIMPLE_EDGE] vectorized | > | PARTITION_ONLY_SHUFFLE [RS_10] | > | Group By Operator [GBY_9] (rows=1 width=72) | > | > Output:["_col0","_col1","_col2","_col3","_col4","_col5","_col6","_col7","_col8"],aggregations:["sum(_col3)","sum(_col0)","count(_col0)","sum(_col5)","sum(_col4)","count(_col1)","sum(_col7)","sum(_col6)","count(_col2)"] > | > | Select Operator [SEL_8] (rows=6 width=232) | > | > Output:["_col0","_col1","_col2","_col3","_col4","_col5","_col6","_col7"] | > | TableScan [TS_0] (rows=6 width=232) | > | default@cbo_test,cbo_test, ACID > table,Tbl:COMPLETE,Col:COMPLETE,Output:["v1","v2","v3"] | > || > ++ > *Query Result* > _c0 _c1 _c2 > 0.0 NaN NaN > *Disable CBO* > ++ > | Explain | > ++ > | Vertex dependency in root stage| > | Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE)| > || > | Stage-0
[jira] [Commented] (HIVE-28258) Use Iceberg semantics for Merge task
[ https://issues.apache.org/jira/browse/HIVE-28258?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17848458#comment-17848458 ] Sourabh Badhya commented on HIVE-28258: --- [~kkasa] , the following task mainly tries to reuse the existing Iceberg readers (IcebergRecordReader) rather than using the file-format readers according to the table format. This way we can use the existing code for handling different file formats (ORC, Parquet, Avro) within Iceberg and avoid writing any custom implementations to handle these file-formats. Additionally, this will help in handling different schemas that Iceberg maintains (the data schema and the delete schema) within Iceberg, and not expose it through public APIs. Custom hacks like changing the file format of the merge task is also removed which was done earlier. The existing tests iceberg_merge_files.q should serve as an example for debugging the merge task used for Iceberg. > Use Iceberg semantics for Merge task > > > Key: HIVE-28258 > URL: https://issues.apache.org/jira/browse/HIVE-28258 > Project: Hive > Issue Type: Improvement > Components: Iceberg integration >Reporter: Sourabh Badhya >Assignee: Sourabh Badhya >Priority: Major > Labels: pull-request-available > > Use Iceberg semantics for Merge task, instead of normal ORC or parquet > readers. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (HIVE-24207) LimitOperator can leverage ObjectCache to bail out quickly
[ https://issues.apache.org/jira/browse/HIVE-24207?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17848432#comment-17848432 ] Sungwoo Park commented on HIVE-24207: - [~abstractdog] Hi, I have a couple of questions of this optimization. 1. An operator tree can contain multiple LimitOperators in general. It seems that this optimization works only if LimitOperator has a single child operator which should be either RS or TerminalOperator. In other words, a vertex should contain a single LimitOperator at most and it should be the last operator before emitting final records. Do you know if this property guaranteed by the Hive compiler? 2. This optimization may not work if speculative execution is enabled or multiple taskattempts are executed in the same LLAP daemon. Or, does this optimization assume no speculative execution? > LimitOperator can leverage ObjectCache to bail out quickly > -- > > Key: HIVE-24207 > URL: https://issues.apache.org/jira/browse/HIVE-24207 > Project: Hive > Issue Type: Improvement >Reporter: Rajesh Balamohan >Assignee: László Bodor >Priority: Major > Labels: pull-request-available > Fix For: 4.0.0-alpha-1 > > Time Spent: 1.5h > Remaining Estimate: 0h > > {noformat} > select ss_sold_date_sk from store_sales, date_dim where date_dim.d_year in > (1998,1998+1,1998+2) and store_sales.ss_sold_date_sk = date_dim.d_date_sk > limit 100; > select distinct ss_sold_date_sk from store_sales, date_dim where > date_dim.d_year in (1998,1998+1,1998+2) and store_sales.ss_sold_date_sk = > date_dim.d_date_sk limit 100; > {noformat} > Queries like the above generate a large number of map tasks. Currently they > don't bail out after generating enough amount of data. > It would be good to make use of ObjectCache & retain the number of records > generated. LimitOperator/VectorLimitOperator can bail out for the later tasks > in the operator's init phase itself. > https://github.com/apache/hive/blob/master/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/VectorLimitOperator.java#L57 > https://github.com/apache/hive/blob/master/ql/src/java/org/apache/hadoop/hive/ql/exec/LimitOperator.java#L58 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (HIVE-28269) Please have regular releases of hive and its docker image
[ https://issues.apache.org/jira/browse/HIVE-28269?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17848323#comment-17848323 ] Raviteja Lokineni commented on HIVE-28269: -- [~ayushtkn] May I ask if there can be a faster release cycle? I’ll try to link all the security tickets here on this one. > Please have regular releases of hive and its docker image > - > > Key: HIVE-28269 > URL: https://issues.apache.org/jira/browse/HIVE-28269 > Project: Hive > Issue Type: Wish >Reporter: Raviteja Lokineni >Priority: Major > > Hi, we as a company are users of Hive metastore and use the docker images. > The latest docker image 4.0.0 has a lot of vulnerabilities. I see most of > them are patched in the mainline code but a release has not been made > available. > Can we/I help in anyway to have regular releases at the very least for the > security patches? if not us then this is request to the hive maintainers to > have regular releases. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28269) Please have regular releases of hive and its docker image
[ https://issues.apache.org/jira/browse/HIVE-28269?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Denys Kuzmenko updated HIVE-28269: -- Priority: Major (was: Blocker) > Please have regular releases of hive and its docker image > - > > Key: HIVE-28269 > URL: https://issues.apache.org/jira/browse/HIVE-28269 > Project: Hive > Issue Type: Task >Reporter: Raviteja Lokineni >Priority: Major > > Hi, we as a company are users of Hive metastore and use the docker images. > The latest docker image 4.0.0 has a lot of vulnerabilities. I see most of > them are patched in the mainline code but a release has not been made > available. > Can we/I help in anyway to have regular releases at the very least for the > security patches? if not us then this is request to the hive maintainers to > have regular releases. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28269) Please have regular releases of hive and its docker image
[ https://issues.apache.org/jira/browse/HIVE-28269?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Denys Kuzmenko updated HIVE-28269: -- Issue Type: Wish (was: Task) > Please have regular releases of hive and its docker image > - > > Key: HIVE-28269 > URL: https://issues.apache.org/jira/browse/HIVE-28269 > Project: Hive > Issue Type: Wish >Reporter: Raviteja Lokineni >Priority: Major > > Hi, we as a company are users of Hive metastore and use the docker images. > The latest docker image 4.0.0 has a lot of vulnerabilities. I see most of > them are patched in the mainline code but a release has not been made > available. > Can we/I help in anyway to have regular releases at the very least for the > security patches? if not us then this is request to the hive maintainers to > have regular releases. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28271) DirectSql fails for AlterPartitions
[ https://issues.apache.org/jira/browse/HIVE-28271?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated HIVE-28271: -- Labels: pull-request-available (was: ) > DirectSql fails for AlterPartitions > --- > > Key: HIVE-28271 > URL: https://issues.apache.org/jira/browse/HIVE-28271 > Project: Hive > Issue Type: Bug >Reporter: Ayush Saxena >Assignee: Ayush Saxena >Priority: Major > Labels: pull-request-available > > It fails at three places: (Misses Database Which Uses CLOB & Missing Boolean > type conversions Checks > *First:* > {noformat} > 2024-05-21T08:50:16,570 WARN [main] metastore.ObjectStore: Falling back to > ORM path due to direct SQL failure (this is not an error): > java.lang.ClassCastException: org.apache.derby.impl.jdbc.EmbedClob cannot be > cast to java.lang.String at > org.apache.hadoop.hive.metastore.ExceptionHandler.newMetaException(ExceptionHandler.java:152) > at org.apache.hadoop.hive.metastore.Batchable.runBatched(Batchable.java:92) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.getParams(DirectSqlUpdatePart.java:748) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateParamTableInBatch(DirectSqlUpdatePart.java:715) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.alterPartitions(DirectSqlUpdatePart.java:636) > at > org.apache.hadoop.hive.metastore.MetaStoreDirectSql.alterPartitions(MetaStoreDirectSql.java:599) > at > org.apache.hadoop.hive.metastore.ObjectStore$20.getSqlResult(ObjectStore.java:5371); > {noformat} > *Second:* > {noformat} > 2024-05-21T09:14:36,808 WARN [main] metastore.ObjectStore: Falling back to > ORM path due to direct SQL failure (this is not an error): > java.lang.ClassCastException: org.apache.derby.impl.jdbc.EmbedClob cannot be > cast to java.lang.String at > org.apache.hadoop.hive.metastore.ExceptionHandler.newMetaException(ExceptionHandler.java:152) > at org.apache.hadoop.hive.metastore.Batchable.runBatched(Batchable.java:92) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateCDInBatch(DirectSqlUpdatePart.java:1228) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateStorageDescriptorInBatch(DirectSqlUpdatePart.java:888) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.alterPartitions(DirectSqlUpdatePart.java:638) > at > org.apache.hadoop.hive.metastore.MetaStoreDirectSql.alterPartitions(MetaStoreDirectSql.java:599) > at > org.apache.hadoop.hive.metastore.ObjectStore$20.getSqlResult(ObjectStore.java:5371);{noformat} > *Third: Missing Boolean check type* > {noformat} > 2024-05-21T09:35:44,063 WARN [main] metastore.ObjectStore: Falling back to > ORM path due to direct SQL failure (this is not an error): > java.sql.BatchUpdateException: A truncation error was encountered trying to > shrink CHAR 'false' to length 1. at > org.apache.hadoop.hive.metastore.ExceptionHandler.newMetaException(ExceptionHandler.java:152) > at org.apache.hadoop.hive.metastore.Batchable.runBatched(Batchable.java:92) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.lambda$updateSDInBatch$16(DirectSqlUpdatePart.java:926) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateWithStatement(DirectSqlUpdatePart.java:656) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateSDInBatch(DirectSqlUpdatePart.java:926) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateStorageDescriptorInBatch(DirectSqlUpdatePart.java:900) > at > org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.alterPartitions(DirectSqlUpdatePart.java:638) > at > org.apache.hadoop.hive.metastore.MetaStoreDirectSql.alterPartitions(MetaStoreDirectSql.java:599) > at > org.apache.hadoop.hive.metastore.ObjectStore$20.getSqlResult(ObjectStore.java:5371); > {noformat} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28271) DirectSql fails for AlterPartitions
Ayush Saxena created HIVE-28271: --- Summary: DirectSql fails for AlterPartitions Key: HIVE-28271 URL: https://issues.apache.org/jira/browse/HIVE-28271 Project: Hive Issue Type: Bug Reporter: Ayush Saxena Assignee: Ayush Saxena It fails at three places: (Misses Database Which Uses CLOB & Missing Boolean type conversions Checks *First:* {noformat} 2024-05-21T08:50:16,570 WARN [main] metastore.ObjectStore: Falling back to ORM path due to direct SQL failure (this is not an error): java.lang.ClassCastException: org.apache.derby.impl.jdbc.EmbedClob cannot be cast to java.lang.String at org.apache.hadoop.hive.metastore.ExceptionHandler.newMetaException(ExceptionHandler.java:152) at org.apache.hadoop.hive.metastore.Batchable.runBatched(Batchable.java:92) at org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.getParams(DirectSqlUpdatePart.java:748) at org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateParamTableInBatch(DirectSqlUpdatePart.java:715) at org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.alterPartitions(DirectSqlUpdatePart.java:636) at org.apache.hadoop.hive.metastore.MetaStoreDirectSql.alterPartitions(MetaStoreDirectSql.java:599) at org.apache.hadoop.hive.metastore.ObjectStore$20.getSqlResult(ObjectStore.java:5371); {noformat} *Second:* {noformat} 2024-05-21T09:14:36,808 WARN [main] metastore.ObjectStore: Falling back to ORM path due to direct SQL failure (this is not an error): java.lang.ClassCastException: org.apache.derby.impl.jdbc.EmbedClob cannot be cast to java.lang.String at org.apache.hadoop.hive.metastore.ExceptionHandler.newMetaException(ExceptionHandler.java:152) at org.apache.hadoop.hive.metastore.Batchable.runBatched(Batchable.java:92) at org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateCDInBatch(DirectSqlUpdatePart.java:1228) at org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateStorageDescriptorInBatch(DirectSqlUpdatePart.java:888) at org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.alterPartitions(DirectSqlUpdatePart.java:638) at org.apache.hadoop.hive.metastore.MetaStoreDirectSql.alterPartitions(MetaStoreDirectSql.java:599) at org.apache.hadoop.hive.metastore.ObjectStore$20.getSqlResult(ObjectStore.java:5371);{noformat} *Third: Missing Boolean check type* {noformat} 2024-05-21T09:35:44,063 WARN [main] metastore.ObjectStore: Falling back to ORM path due to direct SQL failure (this is not an error): java.sql.BatchUpdateException: A truncation error was encountered trying to shrink CHAR 'false' to length 1. at org.apache.hadoop.hive.metastore.ExceptionHandler.newMetaException(ExceptionHandler.java:152) at org.apache.hadoop.hive.metastore.Batchable.runBatched(Batchable.java:92) at org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.lambda$updateSDInBatch$16(DirectSqlUpdatePart.java:926) at org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateWithStatement(DirectSqlUpdatePart.java:656) at org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateSDInBatch(DirectSqlUpdatePart.java:926) at org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.updateStorageDescriptorInBatch(DirectSqlUpdatePart.java:900) at org.apache.hadoop.hive.metastore.DirectSqlUpdatePart.alterPartitions(DirectSqlUpdatePart.java:638) at org.apache.hadoop.hive.metastore.MetaStoreDirectSql.alterPartitions(MetaStoreDirectSql.java:599) at org.apache.hadoop.hive.metastore.ObjectStore$20.getSqlResult(ObjectStore.java:5371); {noformat} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28270) Fix missing partition paths bug on drop_database
[ https://issues.apache.org/jira/browse/HIVE-28270?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated HIVE-28270: -- Labels: pull-request-available (was: ) > Fix missing partition paths bug on drop_database > - > > Key: HIVE-28270 > URL: https://issues.apache.org/jira/browse/HIVE-28270 > Project: Hive > Issue Type: Bug > Components: Hive >Reporter: Wechar >Assignee: Wechar >Priority: Major > Labels: pull-request-available > > In {{HMSHandler#drop_database_core}}, it needs to collect all partition paths > that were not in the subdirectory of the table path, but now it only fetch > the last batch of paths. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (HIVE-28269) Please have regular releases of hive and its docker image
[ https://issues.apache.org/jira/browse/HIVE-28269?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17848287#comment-17848287 ] Ayush Saxena commented on HIVE-28269: - just list the tickets which you want in the next release which fixes the security vulnerabilities, we will have them in 4.0.1 release planned next month. If you intend to fix any which isn't already fixed, create a Jira for that & you can raise a PR to fix them & put hive-4.0.1-must label on those > Please have regular releases of hive and its docker image > - > > Key: HIVE-28269 > URL: https://issues.apache.org/jira/browse/HIVE-28269 > Project: Hive > Issue Type: Task >Reporter: Raviteja Lokineni >Priority: Blocker > > Hi, we as a company are users of Hive metastore and use the docker images. > The latest docker image 4.0.0 has a lot of vulnerabilities. I see most of > them are patched in the mainline code but a release has not been made > available. > Can we/I help in anyway to have regular releases at the very least for the > security patches? if not us then this is request to the hive maintainers to > have regular releases. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28270) Fix missing partition paths bug on drop_database
Wechar created HIVE-28270: - Summary: Fix missing partition paths bug on drop_database Key: HIVE-28270 URL: https://issues.apache.org/jira/browse/HIVE-28270 Project: Hive Issue Type: Bug Components: Hive Reporter: Wechar Assignee: Wechar In {{HMSHandler#drop_database_core}}, it needs to collect all partition paths that were not in the subdirectory of the table path, but now it only fetch the last batch of paths. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28269) Please have regular releases of hive and its docker image
Raviteja Lokineni created HIVE-28269: Summary: Please have regular releases of hive and its docker image Key: HIVE-28269 URL: https://issues.apache.org/jira/browse/HIVE-28269 Project: Hive Issue Type: Task Reporter: Raviteja Lokineni Hi, we as a company are users of Hive metastore and use the docker images. The latest docker image 4.0.0 has a lot of vulnerabilities. I see most of them are patched in the mainline code but a release has not been made available. Can we/I help in anyway to have regular releases at the very least for the security patches? if not us then this is request to the hive maintainers to have regular releases. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (HIVE-28239) Fix bug on HMSHandler#checkLimitNumberOfPartitions
[ https://issues.apache.org/jira/browse/HIVE-28239?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17848213#comment-17848213 ] Denys Kuzmenko commented on HIVE-28239: --- Merged to master thanks for the patch [~wechar]! > Fix bug on HMSHandler#checkLimitNumberOfPartitions > -- > > Key: HIVE-28239 > URL: https://issues.apache.org/jira/browse/HIVE-28239 > Project: Hive > Issue Type: Bug > Components: Hive >Reporter: Wechar >Assignee: Wechar >Priority: Major > Labels: pull-request-available > > {{HMSHandler#checkLimitNumberOfPartitions}} should not compare request size, > which can cause the incorrect limit check. > Assume that HMS configure {{metastore.limit.partition.request}} as 100, the > client calls {{get_partitions_by_filter}} with maxParts as 101, and the > matching partition number is 50, in this case the HMS server should not throw > MetaException by partition limit check. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28239) Fix bug on HMSHandler#checkLimitNumberOfPartitions
[ https://issues.apache.org/jira/browse/HIVE-28239?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Denys Kuzmenko updated HIVE-28239: -- Affects Version/s: 4.0.0 > Fix bug on HMSHandler#checkLimitNumberOfPartitions > -- > > Key: HIVE-28239 > URL: https://issues.apache.org/jira/browse/HIVE-28239 > Project: Hive > Issue Type: Bug > Components: Hive >Affects Versions: 4.0.0 >Reporter: Wechar >Assignee: Wechar >Priority: Major > Labels: pull-request-available > Fix For: 4.1.0 > > > {{HMSHandler#checkLimitNumberOfPartitions}} should not compare request size, > which can cause the incorrect limit check. > Assume that HMS configure {{metastore.limit.partition.request}} as 100, the > client calls {{get_partitions_by_filter}} with maxParts as 101, and the > matching partition number is 50, in this case the HMS server should not throw > MetaException by partition limit check. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (HIVE-28239) Fix bug on HMSHandler#checkLimitNumberOfPartitions
[ https://issues.apache.org/jira/browse/HIVE-28239?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Denys Kuzmenko resolved HIVE-28239. --- Fix Version/s: 4.1.0 Resolution: Fixed > Fix bug on HMSHandler#checkLimitNumberOfPartitions > -- > > Key: HIVE-28239 > URL: https://issues.apache.org/jira/browse/HIVE-28239 > Project: Hive > Issue Type: Bug > Components: Hive >Reporter: Wechar >Assignee: Wechar >Priority: Major > Labels: pull-request-available > Fix For: 4.1.0 > > > {{HMSHandler#checkLimitNumberOfPartitions}} should not compare request size, > which can cause the incorrect limit check. > Assume that HMS configure {{metastore.limit.partition.request}} as 100, the > client calls {{get_partitions_by_filter}} with maxParts as 101, and the > matching partition number is 50, in this case the HMS server should not throw > MetaException by partition limit check. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-26838) Adding support for a new event "Reload event" in the HMS
[ https://issues.apache.org/jira/browse/HIVE-26838?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Manish Maheshwari updated HIVE-26838: - Summary: Adding support for a new event "Reload event" in the HMS (was: Add a new event to improve cache performance in external systems that communicates with HMS.) > Adding support for a new event "Reload event" in the HMS > > > Key: HIVE-26838 > URL: https://issues.apache.org/jira/browse/HIVE-26838 > Project: Hive > Issue Type: Bug > Components: Hive, Standalone Metastore >Reporter: Sai Hemanth Gantasala >Assignee: Sai Hemanth Gantasala >Priority: Major > Labels: pull-request-available > Fix For: 4.0.0-beta-1 > > Time Spent: 3h > Remaining Estimate: 0h > > Adding support for a new event "Reload event" in the HMS (HiveMetaStore). > This event can be used by external services that depend on HMS for metadata > operations to improve its cache performance. In the distributed environment > where there are replicas of an external service (with its own cache in each > of these replicas) talking to HMS for metadata operations, the reload event > can be used to address the cache performance and ensure consistency among all > the replicas for a given table/partition. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28268) Iceberg: Retrieve row count from iceberg SnapshotSummary in case of iceberg.hive.keep.stats=false
Butao Zhang created HIVE-28268: -- Summary: Iceberg: Retrieve row count from iceberg SnapshotSummary in case of iceberg.hive.keep.stats=false Key: HIVE-28268 URL: https://issues.apache.org/jira/browse/HIVE-28268 Project: Hive Issue Type: Task Components: Iceberg integration Reporter: Butao Zhang -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (HIVE-28268) Iceberg: Retrieve row count from iceberg SnapshotSummary in case of iceberg.hive.keep.stats=false
[ https://issues.apache.org/jira/browse/HIVE-28268?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17848151#comment-17848151 ] Butao Zhang commented on HIVE-28268: PR https://github.com/apache/hive/pull/5215 > Iceberg: Retrieve row count from iceberg SnapshotSummary in case of > iceberg.hive.keep.stats=false > - > > Key: HIVE-28268 > URL: https://issues.apache.org/jira/browse/HIVE-28268 > Project: Hive > Issue Type: Task > Components: Iceberg integration >Reporter: Butao Zhang >Priority: Major > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-23993) Handle irrecoverable errors
[ https://issues.apache.org/jira/browse/HIVE-23993?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Smruti Biswal updated HIVE-23993: - Labels: pull-request-available (was: pull pull-request-available) > Handle irrecoverable errors > --- > > Key: HIVE-23993 > URL: https://issues.apache.org/jira/browse/HIVE-23993 > Project: Hive > Issue Type: Task >Reporter: Aasha Medhi >Assignee: Aasha Medhi >Priority: Major > Labels: pull-request-available > Attachments: HIVE-23993.01.patch, HIVE-23993.02.patch, > HIVE-23993.03.patch, HIVE-23993.04.patch, HIVE-23993.05.patch, > HIVE-23993.06.patch, HIVE-23993.07.patch, HIVE-23993.08.patch, Retry Logic > for Replication.pdf > > Time Spent: 2.5h > Remaining Estimate: 0h > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-23993) Handle irrecoverable errors
[ https://issues.apache.org/jira/browse/HIVE-23993?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Smruti Biswal updated HIVE-23993: - Labels: pull pull-request-available (was: pull-request-available) > Handle irrecoverable errors > --- > > Key: HIVE-23993 > URL: https://issues.apache.org/jira/browse/HIVE-23993 > Project: Hive > Issue Type: Task >Reporter: Aasha Medhi >Assignee: Aasha Medhi >Priority: Major > Labels: pull, pull-request-available > Attachments: HIVE-23993.01.patch, HIVE-23993.02.patch, > HIVE-23993.03.patch, HIVE-23993.04.patch, HIVE-23993.05.patch, > HIVE-23993.06.patch, HIVE-23993.07.patch, HIVE-23993.08.patch, Retry Logic > for Replication.pdf > > Time Spent: 2.5h > Remaining Estimate: 0h > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (HIVE-25189) Cache the validWriteIdList in query cache before fetching tables from HMS
[ https://issues.apache.org/jira/browse/HIVE-25189?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17847952#comment-17847952 ] Denys Kuzmenko commented on HIVE-25189: --- [~scarlin], [~kkasa] any ideas if that could be leveraged with HIVE-28238 (cache later when we have types) or could be reverted? > Cache the validWriteIdList in query cache before fetching tables from HMS > - > > Key: HIVE-25189 > URL: https://issues.apache.org/jira/browse/HIVE-25189 > Project: Hive > Issue Type: Improvement > Components: HiveServer2 >Reporter: Steve Carlin >Assignee: Steve Carlin >Priority: Minor > Labels: pull-request-available > Time Spent: 1h 10m > Remaining Estimate: 0h > > For a small performance boost at compile time, we should fetch the > validWriteIdList before fetching the tables. HMS allows these to be batched > together in one call. This will avoid the getTable API from being called > twice, because the first time we call it, we pass in a null for > validWriteIdList. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-26220) Shade & relocate dependencies in hive-exec to avoid conflicting with downstream projects
[ https://issues.apache.org/jira/browse/HIVE-26220?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zhihua Deng updated HIVE-26220: --- Labels: pull-request-available (was: hive-4.0.1-must pull-request-available) > Shade & relocate dependencies in hive-exec to avoid conflicting with > downstream projects > > > Key: HIVE-26220 > URL: https://issues.apache.org/jira/browse/HIVE-26220 > Project: Hive > Issue Type: Improvement >Affects Versions: 4.0.0, 4.0.0-alpha-1 >Reporter: Chao Sun >Priority: Blocker > Labels: pull-request-available > > Currently projects like Spark, Trino/Presto, Iceberg, etc, are depending on > {{hive-exec:core}} which was removed in HIVE-25531. The reason these projects > use {{hive-exec:core}} is because they have the flexibility to exclude, shade > & relocate dependencies in {{hive-exec}} that conflict with the ones they > brought in by themselves. However, with {{hive-exec}} this is no longer > possible, since it is a fat jar that shade those dependencies but do not > relocate many of them. > In order for the downstream projects to consume {{hive-exec}}, we will need > to make sure all the dependencies in {{hive-exec}} are properly shaded and > relocated, so they won't cause conflicts with those from the downstream. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28267) Support merge task functionality for Iceberg delete files
[ https://issues.apache.org/jira/browse/HIVE-28267?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated HIVE-28267: -- Labels: pull-request-available (was: ) > Support merge task functionality for Iceberg delete files > - > > Key: HIVE-28267 > URL: https://issues.apache.org/jira/browse/HIVE-28267 > Project: Hive > Issue Type: Improvement >Reporter: Sourabh Badhya >Assignee: Sourabh Badhya >Priority: Major > Labels: pull-request-available > > Support merge task functionality for Iceberg delete files. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28267) Support merge task functionality for Iceberg delete files
Sourabh Badhya created HIVE-28267: - Summary: Support merge task functionality for Iceberg delete files Key: HIVE-28267 URL: https://issues.apache.org/jira/browse/HIVE-28267 Project: Hive Issue Type: Improvement Reporter: Sourabh Badhya Assignee: Sourabh Badhya Support merge task functionality for Iceberg delete files. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28266) Iceberg: select count(*) from data_files metadata tables gives wrong result
[ https://issues.apache.org/jira/browse/HIVE-28266?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated HIVE-28266: -- Labels: pull-request-available (was: ) > Iceberg: select count(*) from data_files metadata tables gives wrong result > --- > > Key: HIVE-28266 > URL: https://issues.apache.org/jira/browse/HIVE-28266 > Project: Hive > Issue Type: Bug >Reporter: Dmitriy Fingerman >Assignee: Dmitriy Fingerman >Priority: Major > Labels: pull-request-available > > In Hive Iceberg, every table has a corresponding metadata table > "*.data_files" that contains info about the files that contain table's data. > select count(*) from a data_file metadata table returns number of rows in the > data table instead of number of data files from the metadata table. > > {code:java} > CREATE TABLE x (name VARCHAR(50), age TINYINT, num_clicks BIGINT) stored by > iceberg stored as orc TBLPROPERTIES > ('external.table.purge'='true','format-version'='2'); > insert into x values > ('amy', 35, 123412344), > ('adxfvy', 36, 123412534), > ('amsdfyy', 37, 123417234), > ('asafmy', 38, 123412534); > insert into x values > ('amerqwy', 39, 123441234), > ('amyxzcv', 40, 123341234), > ('erweramy', 45, 122341234); > Select * from default.x.data_files; > – Returns 2 records in the output > Select count from default.x.data_files; > – Returns 7 instead of 2 > {code} > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28266) Iceberg: select count(*) from data_files metadata tables gives wrong result
[ https://issues.apache.org/jira/browse/HIVE-28266?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Dmitriy Fingerman updated HIVE-28266: - Summary: Iceberg: select count(*) from data_files metadata tables gives wrong result (was: Iceberg: select count(*) from *.data_files metadata tables gives wrong result) > Iceberg: select count(*) from data_files metadata tables gives wrong result > --- > > Key: HIVE-28266 > URL: https://issues.apache.org/jira/browse/HIVE-28266 > Project: Hive > Issue Type: Bug >Reporter: Dmitriy Fingerman >Assignee: Dmitriy Fingerman >Priority: Major > > In Hive Iceberg, every table has a corresponding metadata table > "*.data_files" that contains info about the files that contain table's data. > select count(*) from a data_file metadata table returns number of rows in the > data table instead of number of data files from the metadata table. > > {code:java} > CREATE TABLE x (name VARCHAR(50), age TINYINT, num_clicks BIGINT) stored by > iceberg stored as orc TBLPROPERTIES > ('external.table.purge'='true','format-version'='2'); > insert into x values > ('amy', 35, 123412344), > ('adxfvy', 36, 123412534), > ('amsdfyy', 37, 123417234), > ('asafmy', 38, 123412534); > insert into x values > ('amerqwy', 39, 123441234), > ('amyxzcv', 40, 123341234), > ('erweramy', 45, 122341234); > Select * from default.x.data_files; > – Returns 2 records in the output > Select count from default.x.data_files; > – Returns 7 instead of 2 > {code} > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28266) Iceberg: select count(*) from *.data_files metadata tables gives wrong result
[ https://issues.apache.org/jira/browse/HIVE-28266?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Dmitriy Fingerman updated HIVE-28266: - Description: In Hive Iceberg, every table has a corresponding metadata table "*.data_files" that contains info about the files that contain table's data. select count(*) from a data_file metadata table returns number of rows in the data table instead of number of data files from the metadata table. {code:java} CREATE TABLE x (name VARCHAR(50), age TINYINT, num_clicks BIGINT) stored by iceberg stored as orc TBLPROPERTIES ('external.table.purge'='true','format-version'='2'); insert into x values ('amy', 35, 123412344), ('adxfvy', 36, 123412534), ('amsdfyy', 37, 123417234), ('asafmy', 38, 123412534); insert into x values ('amerqwy', 39, 123441234), ('amyxzcv', 40, 123341234), ('erweramy', 45, 122341234); Select * from default.x.data_files; – Returns 2 records in the output Select count from default.x.data_files; – Returns 7 instead of 2 {code} was: In Hive Iceberg, every table has a corresponding metadata table "*.data_files" that contains info about the files that contain table's data. select count(*) from a data_file metadata table returns number of rows in the data table instead of number of data files from the metadata table. CREATE TABLE x (name VARCHAR(50), age TINYINT, num_clicks BIGINT) stored by iceberg stored as orc TBLPROPERTIES ('external.table.purge'='true','format-version'='2'); insert into x values ('amy', 35, 123412344), ('adxfvy', 36, 123412534), ('amsdfyy', 37, 123417234), ('asafmy', 38, 123412534); insert into x values ('amerqwy', 39, 123441234), ('amyxzcv', 40, 123341234), ('erweramy', 45, 122341234); Select * from default.x.data_files; -- Returns 2 records in the output Select count(*) from default.x.data_files; -- Returns 7 instead of 2 > Iceberg: select count(*) from *.data_files metadata tables gives wrong result > - > > Key: HIVE-28266 > URL: https://issues.apache.org/jira/browse/HIVE-28266 > Project: Hive > Issue Type: Bug >Reporter: Dmitriy Fingerman >Assignee: Dmitriy Fingerman >Priority: Major > > In Hive Iceberg, every table has a corresponding metadata table > "*.data_files" that contains info about the files that contain table's data. > select count(*) from a data_file metadata table returns number of rows in the > data table instead of number of data files from the metadata table. > > {code:java} > CREATE TABLE x (name VARCHAR(50), age TINYINT, num_clicks BIGINT) stored by > iceberg stored as orc TBLPROPERTIES > ('external.table.purge'='true','format-version'='2'); > insert into x values > ('amy', 35, 123412344), > ('adxfvy', 36, 123412534), > ('amsdfyy', 37, 123417234), > ('asafmy', 38, 123412534); > insert into x values > ('amerqwy', 39, 123441234), > ('amyxzcv', 40, 123341234), > ('erweramy', 45, 122341234); > Select * from default.x.data_files; > – Returns 2 records in the output > Select count from default.x.data_files; > – Returns 7 instead of 2 > {code} > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28266) Iceberg: select count(*) from *.data_files metadata tables gives wrong result
Dmitriy Fingerman created HIVE-28266: Summary: Iceberg: select count(*) from *.data_files metadata tables gives wrong result Key: HIVE-28266 URL: https://issues.apache.org/jira/browse/HIVE-28266 Project: Hive Issue Type: Bug Reporter: Dmitriy Fingerman Assignee: Dmitriy Fingerman In Hive Iceberg, every table has a corresponding metadata table "*.data_files" that contains info about the files that contain table's data. select count(*) from a data_file metadata table returns number of rows in the data table instead of number of data files from the metadata table. CREATE TABLE x (name VARCHAR(50), age TINYINT, num_clicks BIGINT) stored by iceberg stored as orc TBLPROPERTIES ('external.table.purge'='true','format-version'='2'); insert into x values ('amy', 35, 123412344), ('adxfvy', 36, 123412534), ('amsdfyy', 37, 123417234), ('asafmy', 38, 123412534); insert into x values ('amerqwy', 39, 123441234), ('amyxzcv', 40, 123341234), ('erweramy', 45, 122341234); Select * from default.x.data_files; -- Returns 2 records in the output Select count(*) from default.x.data_files; -- Returns 7 instead of 2 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (HIVE-28264) OOM/slow compilation when query contains SELECT clauses with nested expressions
[ https://issues.apache.org/jira/browse/HIVE-28264?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17847000#comment-17847000 ] Alessandro Solimando commented on HIVE-28264: - I guess the problem applies to the respective Calcite rules from which the Hive ones were derived, do you know if that has been addressed there? > OOM/slow compilation when query contains SELECT clauses with nested > expressions > --- > > Key: HIVE-28264 > URL: https://issues.apache.org/jira/browse/HIVE-28264 > Project: Hive > Issue Type: Bug > Components: CBO, HiveServer2 >Affects Versions: 4.0.0 >Reporter: Stamatis Zampetakis >Assignee: Stamatis Zampetakis >Priority: Major > > {code:sql} > CREATE TABLE t0 (`title` string); > SELECT x10 from > (SELECT concat_ws('L10',x9, x9, x9, x9) as x10 from > (SELECT concat_ws('L9',x8, x8, x8, x8) as x9 from > (SELECT concat_ws('L8',x7, x7, x7, x7) as x8 from > (SELECT concat_ws('L7',x6, x6, x6, x6) as x7 from > (SELECT concat_ws('L6',x5, x5, x5, x5) as x6 from > (SELECT concat_ws('L5',x4, x4, x4, x4) as x5 from > (SELECT concat_ws('L4',x3, x3, x3, x3) as x4 from > (SELECT concat_ws('L3',x2, x2, x2, x2) as x3 > from > (SELECT concat_ws('L2',x1, x1, x1, x1) as > x2 from > (SELECT concat_ws('L1',x0, x0, x0, > x0) as x1 from > (SELECT concat_ws('L0',title, > title, title, title) as x0 from t0) t1) t2) t3) t4) t5) t6) t7) t8) t9) t10) t > WHERE x10 = 'Something'; > {code} > The query above fails with OOM when run with the TestMiniLlapLocalCliDriver > and the default max heap size configuration effective for tests (-Xmx2048m). > {noformat} > java.lang.OutOfMemoryError: Java heap space > at java.util.Arrays.copyOf(Arrays.java:3332) > at > java.lang.AbstractStringBuilder.ensureCapacityInternal(AbstractStringBuilder.java:124) > at > java.lang.AbstractStringBuilder.append(AbstractStringBuilder.java:448) > at java.lang.StringBuilder.append(StringBuilder.java:136) > at org.apache.calcite.rex.RexCall.computeDigest(RexCall.java:152) > at org.apache.calcite.rex.RexCall.toString(RexCall.java:165) > at org.apache.calcite.rex.RexCall.appendOperands(RexCall.java:105) > at org.apache.calcite.rex.RexCall.computeDigest(RexCall.java:151) > at org.apache.calcite.rex.RexCall.toString(RexCall.java:165) > at java.lang.String.valueOf(String.java:2994) > at java.lang.StringBuilder.append(StringBuilder.java:131) > at > org.apache.calcite.rel.externalize.RelWriterImpl.explain_(RelWriterImpl.java:90) > at > org.apache.calcite.rel.externalize.RelWriterImpl.done(RelWriterImpl.java:144) > at > org.apache.calcite.rel.AbstractRelNode.explain(AbstractRelNode.java:246) > at > org.apache.calcite.rel.externalize.RelWriterImpl.explainInputs(RelWriterImpl.java:122) > at > org.apache.calcite.rel.externalize.RelWriterImpl.explain_(RelWriterImpl.java:116) > at > org.apache.calcite.rel.externalize.RelWriterImpl.done(RelWriterImpl.java:144) > at > org.apache.calcite.rel.AbstractRelNode.explain(AbstractRelNode.java:246) > at org.apache.calcite.plan.RelOptUtil.toString(RelOptUtil.java:2308) > at org.apache.calcite.plan.RelOptUtil.toString(RelOptUtil.java:2292) > at > org.apache.hadoop.hive.ql.optimizer.calcite.RuleEventLogger.ruleProductionSucceeded(RuleEventLogger.java:73) > at > org.apache.calcite.plan.MulticastRelOptListener.ruleProductionSucceeded(MulticastRelOptListener.java:68) > at > org.apache.calcite.plan.AbstractRelOptPlanner.notifyTransformation(AbstractRelOptPlanner.java:370) > at > org.apache.calcite.plan.hep.HepPlanner.applyTransformationResults(HepPlanner.java:702) > at org.apache.calcite.plan.hep.HepPlanner.applyRule(HepPlanner.java:545) > at > org.apache.calcite.plan.hep.HepPlanner.applyRules(HepPlanner.java:407) > at > org.apache.calcite.plan.hep.HepPlanner.executeInstruction(HepPlanner.java:271) > at > org.apache.calcite.plan.hep.HepInstruction$RuleCollection.execute(HepInstruction.java:74) > at > org.apache.calcite.plan.hep.HepPlanner.executeProgram(HepPlanner.java:202) > at > org.apache.calcite.plan.hep.HepPlanner.findBestExp(HepPlanner.java:189)
[jira] [Commented] (HIVE-28264) OOM/slow compilation when query contains SELECT clauses with nested expressions
[ https://issues.apache.org/jira/browse/HIVE-28264?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17846978#comment-17846978 ] Stamatis Zampetakis commented on HIVE-28264: To understand the problem let's consider a much simpler variation of the query in the description. {code:sql} SELECT x1 from (SELECT concat_ws('L1',x0, x0) as x1 from (SELECT concat_ws('L0',title, title) as x0 from t0) t1) t2; {code} It is easy to see that the SELECT clauses can be merged together leading to the following query. {code:sql} SELECT concat_ws('L1',concat_ws('L0',title, title), concat_ws('L0',title, title)) as x1 from t0; {code} The two queries are equivalent, however they don't contain the same number of {{concat_ws}} calls. The first contains two calls while the second contains three calls and the expression in the SELECT clause is bigger than both of the previous expressions. When the query has nested function calls (CONCAT or anything else) then merging those together leads to bigger expressions. In fact the growth rate of the expression is exponential to the number of its arguments. +Examples:+ When the CONCAT function has two arguments then for each nested level the expression grows by a factor of two. The size of the final expression is (1-2^L)/(1-2) where L is the levels of nesting. When the CONCAT functions has four arguments (as the query in the description) the for each nested level the expression grows by a factor of four. The size of the final expression is (1-4^L)/(1-4) where L is the levels of nesting. There are various optimization rules (eg., HiveFieldTrimmerRule, HiveProjectMergeRule, etc.) that will try to merge expressions together and when this happens in an uncontrolled manner the resulting expression is exponentially big, which can further lead to OOM problems, very slow compilation, etc. Clearly it is not always beneficial to merge expressions together and the aforementioned rules do have some logic in place to avoid this kind of huge expansion. Both rules pass from {{RelOptUtil#pushPastProjectUnlessBloat}} so they can be tuned via the bloat parameter. However, there are also other rules that are affected by this exponential growth problem , such as {{HiveFilterProjectTransposeRule}}, and currently they do not have logic to prevent that. +Before+ {code:sql} SELECT x1 from (SELECT concat_ws('L1',x0, x0) as x1 from (SELECT concat_ws('L0',title, title) as x0 from t0) t1) t2 WHERE x1 = 'Something'; {code} +After+ {code:sql} SELECT x1 from (SELECT concat_ws('L1',x0, x0) as x1 from (SELECT concat_ws('L0',title, title) as x0 from t0 WHERE concat_ws('L1',concat_ws('L0',title, title), concat_ws('L0',title, title)) = 'Something') t1) t2; {code} In this case the exponential growth happens when trying to push the filter down past the projections. A possible solution would be to improve HiveFilterProjectTransposeRule and other rules that may be affected to avoid creating overly complex expressions using a similar bloat configuration parameter. > OOM/slow compilation when query contains SELECT clauses with nested > expressions > --- > > Key: HIVE-28264 > URL: https://issues.apache.org/jira/browse/HIVE-28264 > Project: Hive > Issue Type: Bug > Components: CBO, HiveServer2 >Affects Versions: 4.0.0 >Reporter: Stamatis Zampetakis >Assignee: Stamatis Zampetakis >Priority: Major > > {code:sql} > CREATE TABLE t0 (`title` string); > SELECT x10 from > (SELECT concat_ws('L10',x9, x9, x9, x9) as x10 from > (SELECT concat_ws('L9',x8, x8, x8, x8) as x9 from > (SELECT concat_ws('L8',x7, x7, x7, x7) as x8 from > (SELECT concat_ws('L7',x6, x6, x6, x6) as x7 from > (SELECT concat_ws('L6',x5, x5, x5, x5) as x6 from > (SELECT concat_ws('L5',x4, x4, x4, x4) as x5 from > (SELECT concat_ws('L4',x3, x3, x3, x3) as x4 from > (SELECT concat_ws('L3',x2, x2, x2, x2) as x3 > from > (SELECT concat_ws('L2',x1, x1, x1, x1) as > x2 from > (SELECT concat_ws('L1',x0, x0, x0, > x0) as x1 from > (SELECT concat_ws('L0',title, > title, title, title) as x0 from t0) t1) t2) t3) t4) t5) t6) t7) t8) t9) t10) t > WHERE x10 = 'Something'; > {code} > The query above fails with OOM when run with the TestMiniLlapLocalCliDriver > and the default max heap size configuration effective for tests (-Xmx2048m). &g
[jira] [Updated] (HIVE-28254) CBO (Calcite Return Path): Multiple DISTINCT leads to wrong results
[ https://issues.apache.org/jira/browse/HIVE-28254?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Shohei Okumiya updated HIVE-28254: -- Status: Patch Available (was: Open) > CBO (Calcite Return Path): Multiple DISTINCT leads to wrong results > --- > > Key: HIVE-28254 > URL: https://issues.apache.org/jira/browse/HIVE-28254 > Project: Hive > Issue Type: Bug > Components: CBO >Affects Versions: 4.0.0 >Reporter: Shohei Okumiya >Assignee: Shohei Okumiya >Priority: Major > Labels: hive-4.0.1-must, pull-request-available > > CBO return path can build incorrect GroupByOperator when multiple > aggregations with DISTINCT are involved. > This is an example. > {code:java} > CREATE TABLE test (col1 INT, col2 INT); > INSERT INTO test VALUES (1, 100), (2, 200), (2, 200), (3, 300); > set hive.cbo.returnpath.hiveop=true; > set hive.map.aggr=false; > SELECT > SUM(DISTINCT col1), > COUNT(DISTINCT col1), > SUM(DISTINCT col2), > SUM(col2) > FROM test;{code} > The last column should be 800. But the SUM refers to col1 and the actual > result is 8. > {code:java} > +--+--+--+--+ > | _c0 | _c1 | _c2 | _c3 | > +--+--+--+--+ > | 6 | 3 | 600 | 8 | > +--+--+--+--+ {code} > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28265) Improve the error message for hive.query.timeout.seconds
Shohei Okumiya created HIVE-28265: - Summary: Improve the error message for hive.query.timeout.seconds Key: HIVE-28265 URL: https://issues.apache.org/jira/browse/HIVE-28265 Project: Hive Issue Type: Bug Components: HiveServer2 Affects Versions: 4.0.0 Reporter: Shohei Okumiya Assignee: Shohei Okumiya `hive.query.timeout.seconds` seems to be working correctly, but it always says it timed out in 0 second. {code:java} 0: jdbc:hive2://hive-hiveserver2:1/defaul> set hive.query.timeout.seconds=1s; No rows affected (0.111 seconds) 0: jdbc:hive2://hive-hiveserver2:1/defaul> select count(*) from test; ... Error: Query timed out after 0 seconds (state=,code=0){code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28264) OOM/slow compilation when query contains SELECT clauses with nested expressions
Stamatis Zampetakis created HIVE-28264: -- Summary: OOM/slow compilation when query contains SELECT clauses with nested expressions Key: HIVE-28264 URL: https://issues.apache.org/jira/browse/HIVE-28264 Project: Hive Issue Type: Bug Components: CBO, HiveServer2 Affects Versions: 4.0.0 Reporter: Stamatis Zampetakis Assignee: Stamatis Zampetakis {code:sql} CREATE TABLE t0 (`title` string); SELECT x10 from (SELECT concat_ws('L10',x9, x9, x9, x9) as x10 from (SELECT concat_ws('L9',x8, x8, x8, x8) as x9 from (SELECT concat_ws('L8',x7, x7, x7, x7) as x8 from (SELECT concat_ws('L7',x6, x6, x6, x6) as x7 from (SELECT concat_ws('L6',x5, x5, x5, x5) as x6 from (SELECT concat_ws('L5',x4, x4, x4, x4) as x5 from (SELECT concat_ws('L4',x3, x3, x3, x3) as x4 from (SELECT concat_ws('L3',x2, x2, x2, x2) as x3 from (SELECT concat_ws('L2',x1, x1, x1, x1) as x2 from (SELECT concat_ws('L1',x0, x0, x0, x0) as x1 from (SELECT concat_ws('L0',title, title, title, title) as x0 from t0) t1) t2) t3) t4) t5) t6) t7) t8) t9) t10) t WHERE x10 = 'Something'; {code} The query above fails with OOM when run with the TestMiniLlapLocalCliDriver and the default max heap size configuration effective for tests (-Xmx2048m). {noformat} java.lang.OutOfMemoryError: Java heap space at java.util.Arrays.copyOf(Arrays.java:3332) at java.lang.AbstractStringBuilder.ensureCapacityInternal(AbstractStringBuilder.java:124) at java.lang.AbstractStringBuilder.append(AbstractStringBuilder.java:448) at java.lang.StringBuilder.append(StringBuilder.java:136) at org.apache.calcite.rex.RexCall.computeDigest(RexCall.java:152) at org.apache.calcite.rex.RexCall.toString(RexCall.java:165) at org.apache.calcite.rex.RexCall.appendOperands(RexCall.java:105) at org.apache.calcite.rex.RexCall.computeDigest(RexCall.java:151) at org.apache.calcite.rex.RexCall.toString(RexCall.java:165) at java.lang.String.valueOf(String.java:2994) at java.lang.StringBuilder.append(StringBuilder.java:131) at org.apache.calcite.rel.externalize.RelWriterImpl.explain_(RelWriterImpl.java:90) at org.apache.calcite.rel.externalize.RelWriterImpl.done(RelWriterImpl.java:144) at org.apache.calcite.rel.AbstractRelNode.explain(AbstractRelNode.java:246) at org.apache.calcite.rel.externalize.RelWriterImpl.explainInputs(RelWriterImpl.java:122) at org.apache.calcite.rel.externalize.RelWriterImpl.explain_(RelWriterImpl.java:116) at org.apache.calcite.rel.externalize.RelWriterImpl.done(RelWriterImpl.java:144) at org.apache.calcite.rel.AbstractRelNode.explain(AbstractRelNode.java:246) at org.apache.calcite.plan.RelOptUtil.toString(RelOptUtil.java:2308) at org.apache.calcite.plan.RelOptUtil.toString(RelOptUtil.java:2292) at org.apache.hadoop.hive.ql.optimizer.calcite.RuleEventLogger.ruleProductionSucceeded(RuleEventLogger.java:73) at org.apache.calcite.plan.MulticastRelOptListener.ruleProductionSucceeded(MulticastRelOptListener.java:68) at org.apache.calcite.plan.AbstractRelOptPlanner.notifyTransformation(AbstractRelOptPlanner.java:370) at org.apache.calcite.plan.hep.HepPlanner.applyTransformationResults(HepPlanner.java:702) at org.apache.calcite.plan.hep.HepPlanner.applyRule(HepPlanner.java:545) at org.apache.calcite.plan.hep.HepPlanner.applyRules(HepPlanner.java:407) at org.apache.calcite.plan.hep.HepPlanner.executeInstruction(HepPlanner.java:271) at org.apache.calcite.plan.hep.HepInstruction$RuleCollection.execute(HepInstruction.java:74) at org.apache.calcite.plan.hep.HepPlanner.executeProgram(HepPlanner.java:202) at org.apache.calcite.plan.hep.HepPlanner.findBestExp(HepPlanner.java:189) at org.apache.hadoop.hive.ql.parse.CalcitePlanner$CalcitePlannerAction.executeProgram(CalcitePlanner.java:2452) at org.apache.hadoop.hive.ql.parse.CalcitePlanner$CalcitePlannerAction.executeProgram(CalcitePlanner.java:2411) {noformat} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28263) Metastore scripts : Update query getting stuck when sub-query of in-clause is returning empty results
Taraka Rama Rao Lethavadla created HIVE-28263: - Summary: Metastore scripts : Update query getting stuck when sub-query of in-clause is returning empty results Key: HIVE-28263 URL: https://issues.apache.org/jira/browse/HIVE-28263 Project: Hive Issue Type: Bug Components: Hive Reporter: Taraka Rama Rao Lethavadla As part of fix HIVE-27457 below query is added to [upgrade-4.0.0-alpha-2-to-4.0.0-beta-1.mysql.sql|https://github.com/apache/hive/blob/0e84fe2000c026afd0a49f4e7c7dd5f54fe7b1ec/standalone-metastore/metastore-server/src/main/sql/mysql/upgrade-4.0.0-alpha-2-to-4.0.0-beta-1.mysql.sql#L43] {noformat} UPDATE SERDES SET SERDES.SLIB = "org.apache.hadoop.hive.kudu.KuduSerDe" WHERE SERDE_ID IN ( SELECT SDS.SERDE_ID FROM TBLS INNER JOIN SDS ON TBLS.SD_ID = SDS.SD_ID WHERE TBLS.TBL_ID IN (SELECT TBL_ID FROM TABLE_PARAMS WHERE PARAM_VALUE LIKE '%KuduStorageHandler%') );{noformat} This query is getting hung when sub-query is returning empty results in MySQL {noformat} MariaDB [test]> SELECT TBL_ID FROM table_params WHERE PARAM_VALUE LIKE '%KuduStorageHandler%'; Empty set (0.33 sec) MariaDB [test]> SELECT sds.SERDE_ID FROM tbls LEFT JOIN sds ON tbls.SD_ID = sds.SD_ID WHERE tbls.TBL_ID IN (SELECT TBL_ID FROM table_params WHERE PARAM_VALUE LIKE '%KuduStorageHandler%'); Empty set (0.44 sec) {noformat} And the query kept on running for more than 20 minutes {noformat} MariaDB [test]> UPDATE serdes SET serdes.SLIB = "org.apache.hadoop.hive.kudu.KuduSerDe" WHERE SERDE_ID IN ( SELECT sds.SERDE_ID FROM tbls LEFT JOIN sds ON tbls.SD_ID = sds.SD_ID WHERE tbls.TBL_ID IN (SELECT TBL_ID FROM table_params WHERE PARAM_VALUE LIKE '%KuduStorageHandler%')); ^CCtrl-C -- query killed. Continuing normally. ERROR 1317 (70100): Query execution was interrupted{noformat} The explain extended looks like {noformat} MariaDB [test]> explain extended UPDATE serdes SET serdes.SLIB = "org.apache.hadoop.hive.kudu.KuduSerDe" WHERE SERDE_ID IN ( SELECT sds.SERDE_ID FROM tbls LEFT JOIN sds ON tbls.SD_ID = sds.SD_ID WHERE tbls.TBL_ID IN (SELECT TBL_ID FROM table_params WHERE PARAM_VALUE LIKE '%KuduStorageHandler%')); +--++--++---+--+-+-++--+-+ | id | select_type | table | type | possible_keys | key | key_len | ref | rows | filtered | Extra | +--++--++---+--+-+-++--+-+ | 1 | PRIMARY | serdes | index | NULL | PRIMARY | 8 | NULL | 401267 | 100.00 | Using where | | 2 | DEPENDENT SUBQUERY | tbls | index | PRIMARY,TBLS_N50,TBLS_N49 | TBLS_N50 | 9 | NULL | 50921 | 100.00 | Using index | | 2 | DEPENDENT SUBQUERY | | eq_ref | distinct_key | distinct_key | 8 | func | 1 | 100.00 | | | 2 | DEPENDENT SUBQUERY | sds | eq_ref | PRIMARY | PRIMARY | 8 | test.tbls.SD_ID | 1 | 100.00 | Using where | | 3 | MATERIALIZED | table_params | ALL | PRIMARY,TABLE_PARAMS_N49 | NULL | NULL | NULL | 356593 | 100.00 | Using where | +--++--++---+--+-+-++--+-+ 5 rows in set (0.00 sec){noformat} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28262) Single column use MultiDelimitSerDe parse column error
[ https://issues.apache.org/jira/browse/HIVE-28262?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated HIVE-28262: -- Labels: HiveServer2 pull-request-available (was: HiveServer2) > Single column use MultiDelimitSerDe parse column error > -- > > Key: HIVE-28262 > URL: https://issues.apache.org/jira/browse/HIVE-28262 > Project: Hive > Issue Type: Bug > Components: HiveServer2 >Affects Versions: 3.1.3, 4.1.0 > Environment: Hive version: 3.1.3 >Reporter: Liu Weizheng >Assignee: Liu Weizheng >Priority: Major > Labels: HiveServer2, pull-request-available > Fix For: 4.1.0 > > Attachments: CleanShot 2024-05-16 at 15.13...@2x.png, CleanShot > 2024-05-16 at 15.17...@2x.png > > > ENV: > Hive: 3.1.3/4.1.0 > HDFS: 3.3.1 > -- > Create a text file for external table load,(e.g:/tmp/data): > > {code:java} > 1|@| > 2|@| > 3|@| {code} > > > Create external table: > > {code:java} > CREATE EXTERNAL TABLE IF NOT EXISTS test_split_tmp(`ID` string) ROW FORMAT > SERDE 'org.apache.hadoop.hive.contrib.serde2.MultiDelimitSerDe' WITH > SERDEPROPERTIES('field.delim'='|@|') STORED AS textfile location > '/tmp/test_split_tmp'; {code} > > put text file to external table path: > > {code:java} > hdfs dfs -put /tmp/data /tmp/test_split_tmp {code} > > > query this table and cast column id to long type: > > {code:java} > select UDFToLong(`id`) from test_split_tmp; {code} > *why use UDFToLong function? because it will get NULL result in this > condition,but string type '1' use this function should get type long 1 > result.* > {code:java} > ++ > | id | > ++ > | NULL | > | NULL | > | NULL | > ++ {code} > Therefore, I speculate that there is an issue with the field splitting in > MultiDelimitSerde. > when I debug this issue, I found some problem below: > * org.apache.hadoop.hive.serde2.lazy.LazyStruct#findIndexes > *when fields.length=1 can't find the delimit index* > > {code:java} > private int[] findIndexes(byte[] array, byte[] target) { > if (fields.length <= 1) { // bug > return new int[0]; > } > ... > for (int i = 1; i < indexes.length; i++) { // bug > array = Arrays.copyOfRange(array, indexInNewArray + target.length, > array.length); > indexInNewArray = Bytes.indexOf(array, target); > if (indexInNewArray == -1) { > break; > } > indexes[i] = indexInNewArray + indexes[i - 1] + target.length; > } > return indexes; > }{code} > > * org.apache.hadoop.hive.serde2.lazy.LazyStruct#parseMultiDelimit > *when fields.length=1 can't find the column startPosition* > > {code:java} > public void parseMultiDelimit(byte[] rawRow, byte[] fieldDelimit) { > ... > int[] delimitIndexes = findIndexes(rawRow, fieldDelimit); > ... > if (fields.length > 1 && delimitIndexes[i - 1] != -1) { // bug > int start = delimitIndexes[i - 1] + fieldDelimit.length; > startPosition[i] = start - i * diff; > } else { > startPosition[i] = length + 1; > } > } > Arrays.fill(fieldInited, false); > parsed = true; > }{code} > > > Multi delimit Process: > *Actual:* 1|@| -> 1^A id column start 0 ,next column start 1 > *Expected:* 1|@| -> 1^A id column start 0 ,next column start 2 > > Fix: > # fields.length=1 should find multi delimit index > # fields.length=1 should calculate column start position correct > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (HIVE-28252) AssertionError when using HiveTableScan with a HepPlanner cluster
[ https://issues.apache.org/jira/browse/HIVE-28252?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Stamatis Zampetakis resolved HIVE-28252. Fix Version/s: 4.1.0 Resolution: Fixed Fixed in https://github.com/apache/hive/commit/0e84fe2000c026afd0a49f4e7c7dd5f54fe7b1ec. Thanks for the review [~kkasa]! > AssertionError when using HiveTableScan with a HepPlanner cluster > - > > Key: HIVE-28252 > URL: https://issues.apache.org/jira/browse/HIVE-28252 > Project: Hive > Issue Type: Improvement > Components: CBO, Tests >Affects Versions: 4.0.0 >Reporter: Stamatis Zampetakis >Assignee: Stamatis Zampetakis >Priority: Major > Labels: pull-request-available > Fix For: 4.1.0 > > > The {{HiveTableScan}} operator throws an > [AssertionError|https://github.com/apache/hive/blob/7950967eae9640fcc0aa22f4b6c7906b34281eac/ql/src/java/org/apache/hadoop/hive/ql/optimizer/calcite/reloperators/HiveTableScan.java#L153] > if the operator does not have the {{HiveRelNode.CONVENTION}} set. > The {{HepPlanner}} does not use any > [RelTraitDef|https://github.com/apache/calcite/blob/f854ef5ee480e0ff893b18d27ec67dc381ee2244/core/src/main/java/org/apache/calcite/plan/AbstractRelOptPlanner.java#L276] > so the default [empty traitset for the respective > cluster|https://github.com/apache/calcite/blob/f854ef5ee480e0ff893b18d27ec67dc381ee2244/core/src/main/java/org/apache/calcite/plan/RelOptCluster.java#L99] > is gonna be always empty. > In principle we should not be able to use the {{HiveTableScan}} operator with > {{HepPlanner}}. However, the optimizer heavily uses the {{HepPlanner}} (in > fact more than the {{VolcanoPlanner}} and it is reasonable to wonder how is > this possible given that this assertion is in place. The assertion is > circumvented by creating a cluster from a > [VolcanoPlanner|https://github.com/apache/hive/blob/7950967eae9640fcc0aa22f4b6c7906b34281eac/ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java#L1620] > and then using it in the > [HepPlanner|https://github.com/apache/hive/blob/7950967eae9640fcc0aa22f4b6c7906b34281eac/ql/src/java/org/apache/hadoop/hive/ql/parse/CalcitePlanner.java#L2422]. > > This cluster usage is a bit contrived but does not necessarily need to change > at this stage. > Nevertheless, since the {{HiveTableScan}} operator is suitable to run with > the {{HepPlanner}} the assertion can be relaxed (or removed altogether) to > better reflect the actual usage of the operator, and allow passing a "true" > HepPlanner cluster inside the operator. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28262) Single column use MultiDelimitSerDe parse column error
Liu Weizheng created HIVE-28262: --- Summary: Single column use MultiDelimitSerDe parse column error Key: HIVE-28262 URL: https://issues.apache.org/jira/browse/HIVE-28262 Project: Hive Issue Type: Bug Components: HiveServer2 Affects Versions: 3.1.3, 4.1.0 Environment: Hive version: 3.1.3 Reporter: Liu Weizheng Assignee: Liu Weizheng Fix For: 4.1.0 Attachments: CleanShot 2024-05-16 at 15.13...@2x.png, CleanShot 2024-05-16 at 15.17...@2x.png ENV: Hive: 3.1.3/4.1.0 HDFS: 3.3.1 -- Create a text file for external table load,(e.g:/tmp/data): {code:java} 1|@| 2|@| 3|@| {code} Create external table: {code:java} CREATE EXTERNAL TABLE IF NOT EXISTS test_split_tmp(`ID` string) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.MultiDelimitSerDe' WITH SERDEPROPERTIES('field.delim'='|@|') STORED AS textfile location '/tmp/test_split_tmp'; {code} put text file to external table path: {code:java} hdfs dfs -put /tmp/data /tmp/test_split_tmp {code} query this table and cast column id to long type: {code:java} select UDFToLong(`id`) from test_split_tmp; {code} *why use UDFToLong function? because it will get NULL result in this condition,but string type '1' use this function should get type long 1 result.* {code:java} ++ | id | ++ | NULL | | NULL | | NULL | ++ {code} Therefore, I speculate that there is an issue with the field splitting in MultiDelimitSerde. when I debug this issue, I found some problem below: * org.apache.hadoop.hive.serde2.lazy.LazyStruct#findIndexes *when fields.length=1 can't find the delimit index* {code:java} private int[] findIndexes(byte[] array, byte[] target) { if (fields.length <= 1) { // bug return new int[0]; } ... for (int i = 1; i < indexes.length; i++) { // bug array = Arrays.copyOfRange(array, indexInNewArray + target.length, array.length); indexInNewArray = Bytes.indexOf(array, target); if (indexInNewArray == -1) { break; } indexes[i] = indexInNewArray + indexes[i - 1] + target.length; } return indexes; }{code} * org.apache.hadoop.hive.serde2.lazy.LazyStruct#parseMultiDelimit *when fields.length=1 can't find the column startPosition* {code:java} public void parseMultiDelimit(byte[] rawRow, byte[] fieldDelimit) { ... int[] delimitIndexes = findIndexes(rawRow, fieldDelimit); ... if (fields.length > 1 && delimitIndexes[i - 1] != -1) { // bug int start = delimitIndexes[i - 1] + fieldDelimit.length; startPosition[i] = start - i * diff; } else { startPosition[i] = length + 1; } } Arrays.fill(fieldInited, false); parsed = true; }{code} Multi delimit Process: *Actual:* 1|@| -> 1^A id column start 0 ,next column start 1 *Expected:* 1|@| -> 1^A id column start 0 ,next column start 2 Fix: # fields.length=1 should find multi delimit index # fields.length=1 should calculate column start position correct -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (HIVE-28249) Parquet legacy timezone conversion converts march 1st to 29th feb and fails with not a leap year exception
[ https://issues.apache.org/jira/browse/HIVE-28249?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17846837#comment-17846837 ] Simhadri Govindappa commented on HIVE-28249: Thanks, [~dkuzmenko] and [~zabetak] for the review and all the help :) Change is merged to master. It looks like the jodd authors have acknowledged it as a bug: [https://github.com/oblac/jodd-util/issues/21] . > Parquet legacy timezone conversion converts march 1st to 29th feb and fails > with not a leap year exception > -- > > Key: HIVE-28249 > URL: https://issues.apache.org/jira/browse/HIVE-28249 > Project: Hive > Issue Type: Task >Reporter: Simhadri Govindappa >Assignee: Simhadri Govindappa >Priority: Major > Labels: pull-request-available > > When handling legacy time stamp conversions in parquet,'February 29' year > '200' is an edge case. > This is because, according to this: [https://www.lanl.gov/Caesar/node202.html] > The Julian day for 200 CE/02/29 in the Julian calendar is different from the > Julian day in Gregorian Calendar . > ||Date (BC/AD)||Date (CE)||Julian Day||Julian Day|| > |-| -|(Julian Calendar)|(Gregorian Calendar)| > |200 AD/02/28|200 CE/02/28|1794166|1794167| > |200 AD/02/29|200 CE/02/29|1794167|1794168| > |200 AD/03/01|200 CE/03/01|1794168|1794168| > |300 AD/02/28|300 CE/02/28|1830691|1830691| > |300 AD/02/29|300 CE/02/29|1830692|1830692| > |300 AD/03/01|300 CE/03/01|1830693|1830692| > > * Because of this: > {noformat} > int julianDay = nt.getJulianDay(); {noformat} > returns julian day 1794167 > [https://github.com/apache/hive/blob/master/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/timestamp/NanoTimeUtils.java#L92] > * Later : > {noformat} > Timestamp result = Timestamp.valueOf(formatter.format(date)); {noformat} > _{{{}formatter.format(date{}}})_ returns 29-02-200 as it seems to be using > julian calendar > but _{{Timestamp.valueOf(29-02-200)}}_ seems to be using gregorian calendar > and fails with "not a leap year exception" for 29th Feb 200" > [https://github.com/apache/hive/blob/master/common/src/java/org/apache/hadoop/hive/common/type/TimestampTZUtil.java#L196] > Since hive stores timestamp in UTC, when converting 200 CE/03/01 between > timezones, hive runs into an exception and fails with "not a leap year > exception" for 29th Feb 200 even if the actual record inserted was 200 > CE/03/01 in Asia/Singapore timezone. > > Fullstack trace: > {noformat} > java.lang.RuntimeException: java.io.IOException: > org.apache.parquet.io.ParquetDecodingException: Can not read value at 0 in > block -1 in file > file:/Users/simhadri.govindappa/Documents/apache/hive/itests/qtest/target/localfs/warehouse/test_sgt/sgt000 > at > org.apache.hadoop.hive.ql.exec.FetchTask.executeInner(FetchTask.java:210) > at org.apache.hadoop.hive.ql.exec.FetchTask.execute(FetchTask.java:95) > at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:212) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:154) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:149) > at > org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:185) > at > org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:230) > at > org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:257) > at org.apache.hadoop.hive.cli.CliDriver.processCmd1(CliDriver.java:201) > at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:127) > at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:425) > at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:356) > at > org.apache.hadoop.hive.ql.QTestUtil.executeClientInternal(QTestUtil.java:732) > at org.apache.hadoop.hive.ql.QTestUtil.executeClient(QTestUtil.java:702) > at > org.apache.hadoop.hive.cli.control.CoreCliDriver.runTest(CoreCliDriver.java:116) > at > org.apache.hadoop.hive.cli.control.CliAdapter.runTest(CliAdapter.java:157) > at > org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver(TestMiniLlapLocalCliDriver.java:62) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at &g
[jira] [Resolved] (HIVE-28249) Parquet legacy timezone conversion converts march 1st to 29th feb and fails with not a leap year exception
[ https://issues.apache.org/jira/browse/HIVE-28249?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Simhadri Govindappa resolved HIVE-28249. Fix Version/s: 4.1.0 Resolution: Fixed > Parquet legacy timezone conversion converts march 1st to 29th feb and fails > with not a leap year exception > -- > > Key: HIVE-28249 > URL: https://issues.apache.org/jira/browse/HIVE-28249 > Project: Hive > Issue Type: Task >Reporter: Simhadri Govindappa >Assignee: Simhadri Govindappa >Priority: Major > Labels: pull-request-available > Fix For: 4.1.0 > > > When handling legacy time stamp conversions in parquet,'February 29' year > '200' is an edge case. > This is because, according to this: [https://www.lanl.gov/Caesar/node202.html] > The Julian day for 200 CE/02/29 in the Julian calendar is different from the > Julian day in Gregorian Calendar . > ||Date (BC/AD)||Date (CE)||Julian Day||Julian Day|| > |-| -|(Julian Calendar)|(Gregorian Calendar)| > |200 AD/02/28|200 CE/02/28|1794166|1794167| > |200 AD/02/29|200 CE/02/29|1794167|1794168| > |200 AD/03/01|200 CE/03/01|1794168|1794168| > |300 AD/02/28|300 CE/02/28|1830691|1830691| > |300 AD/02/29|300 CE/02/29|1830692|1830692| > |300 AD/03/01|300 CE/03/01|1830693|1830692| > > * Because of this: > {noformat} > int julianDay = nt.getJulianDay(); {noformat} > returns julian day 1794167 > [https://github.com/apache/hive/blob/master/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/timestamp/NanoTimeUtils.java#L92] > * Later : > {noformat} > Timestamp result = Timestamp.valueOf(formatter.format(date)); {noformat} > _{{{}formatter.format(date{}}})_ returns 29-02-200 as it seems to be using > julian calendar > but _{{Timestamp.valueOf(29-02-200)}}_ seems to be using gregorian calendar > and fails with "not a leap year exception" for 29th Feb 200" > [https://github.com/apache/hive/blob/master/common/src/java/org/apache/hadoop/hive/common/type/TimestampTZUtil.java#L196] > Since hive stores timestamp in UTC, when converting 200 CE/03/01 between > timezones, hive runs into an exception and fails with "not a leap year > exception" for 29th Feb 200 even if the actual record inserted was 200 > CE/03/01 in Asia/Singapore timezone. > > Fullstack trace: > {noformat} > java.lang.RuntimeException: java.io.IOException: > org.apache.parquet.io.ParquetDecodingException: Can not read value at 0 in > block -1 in file > file:/Users/simhadri.govindappa/Documents/apache/hive/itests/qtest/target/localfs/warehouse/test_sgt/sgt000 > at > org.apache.hadoop.hive.ql.exec.FetchTask.executeInner(FetchTask.java:210) > at org.apache.hadoop.hive.ql.exec.FetchTask.execute(FetchTask.java:95) > at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:212) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:154) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:149) > at > org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:185) > at > org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:230) > at > org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:257) > at org.apache.hadoop.hive.cli.CliDriver.processCmd1(CliDriver.java:201) > at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:127) > at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:425) > at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:356) > at > org.apache.hadoop.hive.ql.QTestUtil.executeClientInternal(QTestUtil.java:732) > at org.apache.hadoop.hive.ql.QTestUtil.executeClient(QTestUtil.java:702) > at > org.apache.hadoop.hive.cli.control.CoreCliDriver.runTest(CoreCliDriver.java:116) > at > org.apache.hadoop.hive.cli.control.CliAdapter.runTest(CliAdapter.java:157) > at > org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver(TestMiniLlapLocalCliDriver.java:62) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) > at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable
[jira] [Resolved] (HIVE-28251) HiveSessionImpl init ReaderStream should set Charset with UTF-8
[ https://issues.apache.org/jira/browse/HIVE-28251?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ayush Saxena resolved HIVE-28251. - Resolution: Fixed > HiveSessionImpl init ReaderStream should set Charset with UTF-8 > --- > > Key: HIVE-28251 > URL: https://issues.apache.org/jira/browse/HIVE-28251 > Project: Hive > Issue Type: Improvement >Affects Versions: 3.1.3 >Reporter: xy >Assignee: xy >Priority: Major > Fix For: 4.1.0 > > > Fix some StreamReader not set with UTF8,if we actually default charset not > support Chinese chars such as latin and conf contain Chinese chars,it would > not resolve success,so we need set it as utf8 in StreamReader,we can find all > StreamReader with utf8 charset in other compute framework,such as > Calcite、Hudi and so on. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (HIVE-28251) HiveSessionImpl init ReaderStream should set Charset with UTF-8
[ https://issues.apache.org/jira/browse/HIVE-28251?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17846748#comment-17846748 ] Ayush Saxena commented on HIVE-28251: - Committed to master. Thanx [~xuzifu] for the contribution!!! > HiveSessionImpl init ReaderStream should set Charset with UTF-8 > --- > > Key: HIVE-28251 > URL: https://issues.apache.org/jira/browse/HIVE-28251 > Project: Hive > Issue Type: Improvement >Affects Versions: 3.1.3 >Reporter: xy >Assignee: xy >Priority: Major > Fix For: 4.1.0 > > > Fix some StreamReader not set with UTF8,if we actually default charset not > support Chinese chars such as latin and conf contain Chinese chars,it would > not resolve success,so we need set it as utf8 in StreamReader,we can find all > StreamReader with utf8 charset in other compute framework,such as > Calcite、Hudi and so on. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28251) HiveSessionImpl init ReaderStream should set Charset with UTF-8
[ https://issues.apache.org/jira/browse/HIVE-28251?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ayush Saxena updated HIVE-28251: Summary: HiveSessionImpl init ReaderStream should set Charset with UTF-8 (was: HiveSessionImpl init ReaderStream set Charset with UTF-8) > HiveSessionImpl init ReaderStream should set Charset with UTF-8 > --- > > Key: HIVE-28251 > URL: https://issues.apache.org/jira/browse/HIVE-28251 > Project: Hive > Issue Type: Improvement >Affects Versions: 3.1.3 >Reporter: xy >Assignee: xy >Priority: Major > Fix For: 4.1.0 > > > Fix some StreamReader not set with UTF8,if we actually default charset not > support Chinese chars such as latin and conf contain Chinese chars,it would > not resolve success,so we need set it as utf8 in StreamReader,we can find all > StreamReader with utf8 charset in other compute framework,such as > Calcite、Hudi and so on. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-27361) Add table level comments to Hive tables in SYS & Information schema databases to make them self explanatory
[ https://issues.apache.org/jira/browse/HIVE-27361?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Smruti Biswal updated HIVE-27361: - Labels: pull-request-available (was: ) > Add table level comments to Hive tables in SYS & Information schema databases > to make them self explanatory > --- > > Key: HIVE-27361 > URL: https://issues.apache.org/jira/browse/HIVE-27361 > Project: Hive > Issue Type: Improvement > Components: Hive >Reporter: Taraka Rama Rao Lethavadla >Assignee: Smruti Biswal >Priority: Minor > Labels: pull-request-available > > All the backend database tables are available to query in Hive as external > tables in SYS and information_schema databases. We can add a table level > comment to each of the table while creating them in Hive, so that users can > do desc formatted table_name to check the information about that table -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-20220) Incorrect result when hive.groupby.skewindata is enabled
[ https://issues.apache.org/jira/browse/HIVE-20220?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jiayi Liu updated HIVE-20220: - Description: hive.groupby.skewindata makes use of rand UDF to randomly distribute grouped by keys to the reducers and hence avoids overloading a single reducer when there is a skew in data. This random distribution of keys is buggy when the reducer fails to fetch the mapper output due to a faulty datanode or any other reason. When reducer finds that it can't fetch mapper output, it sends a signal to Application Master to reattempt the corresponding map task. The reattempted map task will now get the different random value from rand function and hence the keys that gets distributed now to the reducer will not be same as the previous run. *Steps to reproduce:* create table test(id int); insert into test values (1),(2),(2),(3),(3),(3),(4),(4),(4),(4),(5),(5),(5),(5),(5),(6),(6),(6),(6),(6),(6),(7),(7),(7),(7),(7),(7),(7),(7),(8),(8),(8),(8),(8),(8),(8),(8),(9),(9),(9),(9),(9),(9),(9),(9),(9); SET hive.groupby.skewindata=true; SET mapreduce.reduce.reduces=2; //Add a debug port for reducer select count(1) from test group by id; //Remove mapper's intermediate output file when map stage is completed and one out of 2 reduce tasks is completed and then continue the run. This causes 2nd reducer to send event to Application Master to rerun the map task. The following is the expected result. 1 2 3 4 5 6 8 8 9 But you may get different result due to a different value returned by the rand function in the second run causing different distribution of keys. This needs to be fixed such that the mapper distributes the same keys even if it is reattempted multiple times. was: hive.groupby.skewindata makes use of rand UDF to randomly distribute grouped by keys to the reducers and hence avoids overloading a single reducer when there is a skew in data. This random distribution of keys is buggy when the reducer fails to fetch the mapper output due to a faulty datanode or any other reason. When reducer finds that it can't fetch mapper output, it sends a signal to Application Master to reattempt the corresponding map task. The reattempted map task will now get the different random value from rand function and hence the keys that gets distributed now to the reducer will not be same as the previous run. *Steps to reproduce:* create table test(id int); insert into test values (1),(2),(2),(3),(3),(3),(4),(4),(4),(4),(5),(5),(5),(5),(5),(6),(6),(6),(6),(6),(6),(7),(7),(7),(7),(7),(7),(7),(7),(8),(8),(8),(8),(8),(8),(8),(8),(9),(9),(9),(9),(9),(9),(9),(9),(9); SET hive.groupby.skewindata=true; SET mapreduce.reduce.reduces=2; //Add a debug port for reducer select count(1) from test group by id; //Remove mapper's intermediate output file when map stage is completed and one out of 2 reduce tasks is completed and then continue the run. This causes 2nd reducer to send event to Application Master to rerun the map task. The following is the expected result. 1 2 3 4 5 6 8 8 9 But you may get different result due to a different value returned by the rand function in the second run causing different distribution of keys. This needs to be fixed such that the mapper distributes the same keys even if it is reattempted multiple times. > Incorrect result when hive.groupby.skewindata is enabled > > > Key: HIVE-20220 > URL: https://issues.apache.org/jira/browse/HIVE-20220 > Project: Hive > Issue Type: Bug > Components: Query Processor >Affects Versions: 3.0.0 >Reporter: Ganesha Shreedhara >Assignee: Ganesha Shreedhara >Priority: Major > Attachments: HIVE-20220.2.patch, HIVE-20220.patch > > > hive.groupby.skewindata makes use of rand UDF to randomly distribute grouped > by keys to the reducers and hence avoids overloading a single reducer when > there is a skew in data. > This random distribution of keys is buggy when the reducer fails to fetch the > mapper output due to a faulty datanode or any other reason. When reducer > finds that it can't fetch mapper output, it sends a signal to Application > Master to reattempt the corresponding map task. The reattempted map task will > now get the different random value from rand function and hence the keys that > gets distributed now to the reducer will not be same as the previous run. > > *Steps to reproduce:* > create table test(id int); > insert into test values > (1),(2),(2),(3),(3),(3),(4),(4),(4),(4),(5),(5),(5),(5),(5),(6),(6),(6),(6),(6),(6),(7),(7),(7),(7),(7),(7),(7),(7),(8),(8),(8),(8),(8),(8),(8),(8),(9),(9),(9),(9),(9),(9),(9),(9),(9); > SET hive.groupby.skewindata
[jira] [Comment Edited] (HIVE-28261) Update Hive version in Docker README
[ https://issues.apache.org/jira/browse/HIVE-28261?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17846550#comment-17846550 ] Zhihua Deng edited comment on HIVE-28261 at 5/15/24 8:49 AM: - Thank you for the PR Mert! Feel free to change the assignee to you if the account is ready. was (Author: dengzh): Thank you for the PR Mert!. Feel free to change the assignee to you if the account is ready. > Update Hive version in Docker README > > > Key: HIVE-28261 > URL: https://issues.apache.org/jira/browse/HIVE-28261 > Project: Hive > Issue Type: Improvement >Reporter: Zhihua Deng >Priority: Major > Fix For: 4.1.0 > > > Add the updates on the page to the readme as quickstart shows. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (HIVE-28261) Update Hive version in Docker README
[ https://issues.apache.org/jira/browse/HIVE-28261?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zhihua Deng resolved HIVE-28261. Fix Version/s: 4.1.0 Resolution: Fixed Thank you for the PR Mert!. Feel free to change the assignee to you if the account is ready. > Update Hive version in Docker README > > > Key: HIVE-28261 > URL: https://issues.apache.org/jira/browse/HIVE-28261 > Project: Hive > Issue Type: Improvement >Reporter: Zhihua Deng >Priority: Major > Fix For: 4.1.0 > > > Add the updates on the page to the readme as quickstart shows. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28261) Update Hive version in Docker README
[ https://issues.apache.org/jira/browse/HIVE-28261?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zhihua Deng updated HIVE-28261: --- Description: Add the updates on the page to the readme as quickstart shows. > Update Hive version in Docker README > > > Key: HIVE-28261 > URL: https://issues.apache.org/jira/browse/HIVE-28261 > Project: Hive > Issue Type: Improvement >Reporter: Zhihua Deng >Priority: Major > > Add the updates on the page to the readme as quickstart shows. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28261) Update Hive version in Docker README
Zhihua Deng created HIVE-28261: -- Summary: Update Hive version in Docker README Key: HIVE-28261 URL: https://issues.apache.org/jira/browse/HIVE-28261 Project: Hive Issue Type: Improvement Reporter: Zhihua Deng -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28098) Fails to copy empty column statistics of materialized CTE
[ https://issues.apache.org/jira/browse/HIVE-28098?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zhihua Deng updated HIVE-28098: --- Labels: hive-4.0.1-must pull-request-available (was: pull-request-available) > Fails to copy empty column statistics of materialized CTE > - > > Key: HIVE-28098 > URL: https://issues.apache.org/jira/browse/HIVE-28098 > Project: Hive > Issue Type: Bug > Components: Query Planning >Reporter: Shohei Okumiya >Assignee: Shohei Okumiya >Priority: Major > Labels: hive-4.0.1-must, pull-request-available > Fix For: 4.1.0 > > > HIVE-28080 introduced the optimization of materialized CTEs, but it turned > out that it failed when statistics were empty. > This query reproduces the issue. > {code:java} > set hive.stats.autogather=false; > CREATE TABLE src_no_stats AS SELECT '123' as key, 'val123' as value UNION ALL > SELECT '9' as key, 'val9' as value; > set hive.optimize.cte.materialize.threshold=2; > set hive.optimize.cte.materialize.full.aggregate.only=false; > EXPLAIN WITH materialized_cte1 AS ( > SELECT * FROM src_no_stats > ), > materialized_cte2 AS ( > SELECT a.key > FROM materialized_cte1 a > JOIN materialized_cte1 b ON (a.key = b.key) > ) > SELECT a.key > FROM materialized_cte2 a > JOIN materialized_cte2 b ON (a.key = b.key); {code} > It throws an error. > {code:java} > Error: Error while compiling statement: FAILED: IllegalStateException The > size of col stats must be equal to that of schema. Stats = [], Schema = [key] > (state=42000,code=4) {code} > Attaching a debugger, FSO of materialized_cte2 has empty stats as > JoinOperator loses stats. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-27847) Prevent query Failures on Numeric <-> Timestamp
[ https://issues.apache.org/jira/browse/HIVE-27847?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zhihua Deng updated HIVE-27847: --- Labels: hive-4.0.1-must pull-request-available (was: pull-request-available) > Prevent query Failures on Numeric <-> Timestamp > > > Key: HIVE-27847 > URL: https://issues.apache.org/jira/browse/HIVE-27847 > Project: Hive > Issue Type: Bug > Components: Hive >Affects Versions: 4.0.0-alpha-1, 4.0.0-alpha-2 > Environment: master > 4.0.0-alpha-1 >Reporter: Basapuram Kumar >Assignee: Basapuram Kumar >Priority: Major > Labels: hive-4.0.1-must, pull-request-available > Attachments: HIVE-27847.patch > > > In Master/4.0.0-alpha-1 branches, performing the Numeric to Timestamp > conversion, its failing with the error as > "{color:#de350b}org.apache.hadoop.hive.ql.exec.UDFArgumentException: Casting > NUMERIC types to TIMESTAMP is prohibited > (hive.strict.timestamp.conversion){color}" . > > *Repro steps.* > # Sample data > {noformat} > $ hdfs dfs -cat /tmp/tc/t.csv > 1653209895687,2022-05-22T15:58:15.931+07:00 > 1653209938316,2022-05-22T15:58:58.490+07:00 > 1653209962021,2022-05-22T15:59:22.191+07:00 > 1653210021993,2022-05-22T16:00:22.174+07:00 > 1653209890524,2022-05-22T15:58:10.724+07:00 > 1653210095382,2022-05-22T16:01:35.775+07:00 > 1653210044308,2022-05-22T16:00:44.683+07:00 > 1653210098546,2022-05-22T16:01:38.886+07:00 > 1653210012220,2022-05-22T16:00:12.394+07:00 > 165321376,2022-05-22T16:00:00.622+07:00{noformat} > # table with above data [1] > {noformat} > create external table test_ts_conv(begin string, ts string) row format > delimited fields terminated by ',' stored as TEXTFILE LOCATION '/tmp/tc/'; > desc test_ts_conv; > | col_name | data_type | comment | > +---++--+ > | begin | string | | > | ts | string | | > +---++--+{noformat} > # Create table with CTAS > {noformat} > 0: jdbc:hive2://char1000.sre.iti.acceldata.de> set > hive.strict.timestamp.conversion; > +-+ > | set | > +-+ > | hive.strict.timestamp.conversion=true | > +-+ > set to false > 0: jdbc:hive2://char1000.sre.iti.acceldata.de> set > hive.strict.timestamp.conversion=false; > +-+ > | set | > +-+ > | hive.strict.timestamp.conversion=false | > +-+ > #Query: > 0: jdbc:hive2://char1000.sre.iti.acceldata.de> > CREATE TABLE t_date > AS > select > CAST( CAST( `begin` AS BIGINT) / 1000 AS TIMESTAMP ) `begin`, > CAST( > DATE_FORMAT(CAST(regexp_replace(`ts`,'(\\d{4})-(\\d{2})-(\\d{2})T(\\d{2}):(\\d{2}):(\\d{2}).(\\d{3})\\+(\\d{2}):(\\d{2})','$1-$2-$3 > $4:$5:$6.$7') AS TIMESTAMP ),'MMdd') as BIGINT ) `par_key` > FROM test_ts_conv;{noformat} > Error: > {code:java} > Caused by: org.apache.hadoop.hive.ql.exec.UDFArgumentException: Casting > NUMERIC types to TIMESTAMP is prohibited (hive.strict.timestamp.conversion) > at > org.apache.hadoop.hive.ql.udf.generic.GenericUDFTimestamp.initialize(GenericUDFTimestamp.java:91) > at > org.apache.hadoop.hive.ql.udf.generic.GenericUDF.initializeAndFoldConstants(GenericUDF.java:149) > at > org.apache.hadoop.hive.ql.exec.ExprNodeGenericFuncEvaluator.initialize(ExprNodeGenericFuncEvaluator.java:184) > at > org.apache.hadoop.hive.ql.exec.Operator.initEvaluators(Operator.java:1073) > at > org.apache.hadoop.hive.ql.exec.Operator.initEvaluatorsAndReturnStruct(Operator.java:1099) > at > org.apache.hadoop.hive.ql.exec.SelectOperator.initializeOp(SelectOperator.java:74) > at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:360) > at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:549) > at > org.apache.hadoop.hive.ql.exec.Operator.initializeChildren(Operator.java:503) > at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:369) > at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:549) > at > org.apache.hadoop.hive.ql.exec.Operator.initializeChildren(Operator.java:503) > at org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:369) > at > org.apache.hadoop.hive.ql.exec.MapOperator.initializeMapOperator(MapOperator.java:508) > at > org.apache.hadoop.hive.ql.exec.tez.MapRecordProcessor.init(MapRecordProcessor.java:314) > ... 17 more {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28258) Use Iceberg semantics for Merge task
[ https://issues.apache.org/jira/browse/HIVE-28258?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated HIVE-28258: -- Labels: pull-request-available (was: ) > Use Iceberg semantics for Merge task > > > Key: HIVE-28258 > URL: https://issues.apache.org/jira/browse/HIVE-28258 > Project: Hive > Issue Type: Improvement > Components: Iceberg integration >Reporter: Sourabh Badhya >Assignee: Sourabh Badhya >Priority: Major > Labels: pull-request-available > > Use Iceberg semantics for Merge task, instead of normal ORC or parquet > readers. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28260) CreateTableEvent wrongly skips authorizing DFS_URI for managed table
[ https://issues.apache.org/jira/browse/HIVE-28260?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ayush Saxena updated HIVE-28260: Labels: hive-4.0.1-must pull-request-available (was: pull-request-available) > CreateTableEvent wrongly skips authorizing DFS_URI for managed table > - > > Key: HIVE-28260 > URL: https://issues.apache.org/jira/browse/HIVE-28260 > Project: Hive > Issue Type: Bug >Reporter: Ayush Saxena >Assignee: Ayush Saxena >Priority: Major > Labels: hive-4.0.1-must, pull-request-available > Fix For: 4.1.0 > > > HIVE-27525 eased out permissions for external table but it wrongly eased out > for managed tables as well by wrong check for managed tables -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (HIVE-28260) CreateTableEvent wrongly skips authorizing DFS_URI for managed table
[ https://issues.apache.org/jira/browse/HIVE-28260?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ayush Saxena resolved HIVE-28260. - Fix Version/s: 4.1.0 Resolution: Fixed > CreateTableEvent wrongly skips authorizing DFS_URI for managed table > - > > Key: HIVE-28260 > URL: https://issues.apache.org/jira/browse/HIVE-28260 > Project: Hive > Issue Type: Bug >Reporter: Ayush Saxena >Assignee: Ayush Saxena >Priority: Major > Labels: pull-request-available > Fix For: 4.1.0 > > > HIVE-27525 eased out permissions for external table but it wrongly eased out > for managed tables as well by wrong check for managed tables -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (HIVE-28260) CreateTableEvent wrongly skips authorizing DFS_URI for managed table
[ https://issues.apache.org/jira/browse/HIVE-28260?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17846501#comment-17846501 ] Ayush Saxena commented on HIVE-28260: - Committed to master. Thanx [~sbadhya] & [~aturoczy] for the review!!! > CreateTableEvent wrongly skips authorizing DFS_URI for managed table > - > > Key: HIVE-28260 > URL: https://issues.apache.org/jira/browse/HIVE-28260 > Project: Hive > Issue Type: Bug >Reporter: Ayush Saxena >Assignee: Ayush Saxena >Priority: Major > Labels: pull-request-available > > HIVE-27525 eased out permissions for external table but it wrongly eased out > for managed tables as well by wrong check for managed tables -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28260) CreateTableEvent wrongly skips authorizing DFS_URI for managed table
[ https://issues.apache.org/jira/browse/HIVE-28260?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated HIVE-28260: -- Labels: pull-request-available (was: ) > CreateTableEvent wrongly skips authorizing DFS_URI for managed table > - > > Key: HIVE-28260 > URL: https://issues.apache.org/jira/browse/HIVE-28260 > Project: Hive > Issue Type: Bug >Reporter: Ayush Saxena >Assignee: Ayush Saxena >Priority: Major > Labels: pull-request-available > > HIVE-27525 eased out permissions for external table but it wrongly eased out > for managed tables as well by wrong check for managed tables -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28260) CreateTableEvent wrongly skips authorizing DFS_URI for managed table
Ayush Saxena created HIVE-28260: --- Summary: CreateTableEvent wrongly skips authorizing DFS_URI for managed table Key: HIVE-28260 URL: https://issues.apache.org/jira/browse/HIVE-28260 Project: Hive Issue Type: Bug Reporter: Ayush Saxena Assignee: Ayush Saxena HIVE-27525 eased out permissions for external table but it wrongly eased out for managed tables as well by wrong check for managed tables -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (HIVE-28259) Common table expression detection and rewrites using CBO
[ https://issues.apache.org/jira/browse/HIVE-28259?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated HIVE-28259: -- Labels: pull-request-available (was: ) > Common table expression detection and rewrites using CBO > > > Key: HIVE-28259 > URL: https://issues.apache.org/jira/browse/HIVE-28259 > Project: Hive > Issue Type: New Feature > Components: CBO >Reporter: Stamatis Zampetakis >Assignee: Stamatis Zampetakis >Priority: Major > Labels: pull-request-available > > Hive already provides the means to detect and exploit CTEs via the > {{SharedWorkOptimizer}}. The {{SharedWorkOptimizer}} relies on a series of > heuristic transformations of the physical plan ({{Operator}} DAG) that apply > towards the end of the planning process. The optimizer is quite powerful and > offers various properties ({{hive.optimize.shared.work*}}) through which its > behavior can be fine tuned by users but has also a few drawbacks: > * not cost-based > * limited customization > * complex implementation > This ticket aims to leverage CBO for detecting and exploiting common table > expressions (CTE) in queries in an attempt to alleviate some of the > shortcomings of the {{SharedWorkOptimizer}} and open the road for more > powerful transformations. > The initial work focuses on establishing the general design and the main APIs > for CBO based CTE transformations. > The main idea is to model CTEs as materialized views (MVs) and rely on > existing MV rewrite logic to incorporate them in the plan in a cost-based > fashion as a new CBO planning phase. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Work started] (HIVE-28259) Common table expression detection and rewrites using CBO
[ https://issues.apache.org/jira/browse/HIVE-28259?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Work on HIVE-28259 started by Stamatis Zampetakis. -- > Common table expression detection and rewrites using CBO > > > Key: HIVE-28259 > URL: https://issues.apache.org/jira/browse/HIVE-28259 > Project: Hive > Issue Type: New Feature > Components: CBO >Reporter: Stamatis Zampetakis >Assignee: Stamatis Zampetakis >Priority: Major > > Hive already provides the means to detect and exploit CTEs via the > {{SharedWorkOptimizer}}. The {{SharedWorkOptimizer}} relies on a series of > heuristic transformations of the physical plan ({{Operator}} DAG) that apply > towards the end of the planning process. The optimizer is quite powerful and > offers various properties ({{hive.optimize.shared.work*}}) through which its > behavior can be fine tuned by users but has also a few drawbacks: > * not cost-based > * limited customization > * complex implementation > This ticket aims to leverage CBO for detecting and exploiting common table > expressions (CTE) in queries in an attempt to alleviate some of the > shortcomings of the {{SharedWorkOptimizer}} and open the road for more > powerful transformations. > The initial work focuses on establishing the general design and the main APIs > for CBO based CTE transformations. > The main idea is to model CTEs as materialized views (MVs) and rely on > existing MV rewrite logic to incorporate them in the plan in a cost-based > fashion as a new CBO planning phase. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (HIVE-28259) Common table expression detection and rewrites using CBO
Stamatis Zampetakis created HIVE-28259: -- Summary: Common table expression detection and rewrites using CBO Key: HIVE-28259 URL: https://issues.apache.org/jira/browse/HIVE-28259 Project: Hive Issue Type: New Feature Components: CBO Reporter: Stamatis Zampetakis Assignee: Stamatis Zampetakis Hive already provides the means to detect and exploit CTEs via the {{SharedWorkOptimizer}}. The {{SharedWorkOptimizer}} relies on a series of heuristic transformations of the physical plan ({{Operator}} DAG) that apply towards the end of the planning process. The optimizer is quite powerful and offers various properties ({{hive.optimize.shared.work*}}) through which its behavior can be fine tuned by users but has also a few drawbacks: * not cost-based * limited customization * complex implementation This ticket aims to leverage CBO for detecting and exploiting common table expressions (CTE) in queries in an attempt to alleviate some of the shortcomings of the {{SharedWorkOptimizer}} and open the road for more powerful transformations. The initial work focuses on establishing the general design and the main APIs for CBO based CTE transformations. The main idea is to model CTEs as materialized views (MVs) and rely on existing MV rewrite logic to incorporate them in the plan in a cost-based fashion as a new CBO planning phase. -- This message was sent by Atlassian Jira (v8.20.10#820010)