rdblue commented on a change in pull request #1870: URL: https://github.com/apache/iceberg/pull/1870#discussion_r654051672
########## File path: core/src/main/java/org/apache/iceberg/jdbc/JdbcCatalog.java ########## @@ -0,0 +1,388 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, + * software distributed under the License is distributed on an + * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY + * KIND, either express or implied. See the License for the + * specific language governing permissions and limitations + * under the License. + */ + +package org.apache.iceberg.jdbc; + +import java.io.Closeable; +import java.sql.DatabaseMetaData; +import java.sql.PreparedStatement; +import java.sql.ResultSet; +import java.sql.SQLException; +import java.sql.SQLIntegrityConstraintViolationException; +import java.sql.SQLNonTransientConnectionException; +import java.sql.SQLTimeoutException; +import java.sql.SQLTransientConnectionException; +import java.util.Arrays; +import java.util.List; +import java.util.Map; +import java.util.Set; +import java.util.stream.Collectors; +import org.apache.hadoop.conf.Configurable; +import org.apache.hadoop.conf.Configuration; +import org.apache.iceberg.BaseMetastoreCatalog; +import org.apache.iceberg.CatalogProperties; +import org.apache.iceberg.CatalogUtil; +import org.apache.iceberg.TableMetadata; +import org.apache.iceberg.TableOperations; +import org.apache.iceberg.catalog.Namespace; +import org.apache.iceberg.catalog.SupportsNamespaces; +import org.apache.iceberg.catalog.TableIdentifier; +import org.apache.iceberg.exceptions.AlreadyExistsException; +import org.apache.iceberg.exceptions.NamespaceNotEmptyException; +import org.apache.iceberg.exceptions.NoSuchNamespaceException; +import org.apache.iceberg.exceptions.NoSuchTableException; +import org.apache.iceberg.hadoop.HadoopFileIO; +import org.apache.iceberg.io.FileIO; +import org.apache.iceberg.relocated.com.google.common.base.Joiner; +import org.apache.iceberg.relocated.com.google.common.base.Preconditions; +import org.apache.iceberg.relocated.com.google.common.collect.ImmutableMap; +import org.apache.iceberg.relocated.com.google.common.collect.Lists; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +public class JdbcCatalog extends BaseMetastoreCatalog implements Configurable, SupportsNamespaces, Closeable { + + public static final String PROPERTY_PREFIX = "connection.parameter."; + private static final Logger LOG = LoggerFactory.getLogger(JdbcCatalog.class); + private static final Joiner SLASH = Joiner.on("/"); + + private FileIO io; + private String catalogName = "jdbc"; + private String warehouseLocation; + private Configuration conf; + private JdbcClientPool connections; + + public JdbcCatalog() { + } + + @Override + public void initialize(String name, Map<String, String> properties) { + String uri = properties.get(CatalogProperties.URI); + Preconditions.checkNotNull(uri, "JDBC connection URI is required"); + + String warehouse = properties.get(CatalogProperties.WAREHOUSE_LOCATION); + Preconditions.checkNotNull(warehouse, "JDBC warehouse location is required"); + this.warehouseLocation = warehouse.replaceAll("/$", ""); + + if (name != null) { + this.catalogName = name; + } + + String fileIOImpl = properties.get(CatalogProperties.FILE_IO_IMPL); + this.io = fileIOImpl == null ? new HadoopFileIO(conf) : CatalogUtil.loadFileIO(fileIOImpl, properties, conf); + + try { + LOG.debug("Connecting to Jdbc database {}", properties.get(CatalogProperties.URI)); + connections = new JdbcClientPool(uri, properties); + initializeCatalogTables(); + } catch (SQLTimeoutException e) { + throw new UncheckedSQLException(e, "Cannot initialize jdbc catalog: Query timed out"); + } catch (SQLTransientConnectionException | SQLNonTransientConnectionException e) { + throw new UncheckedSQLException(e, "Cannot initialize jdbc catalog: Connection failed"); + } catch (SQLException e) { + throw new UncheckedSQLException(e, "Cannot initialize jdbc catalog"); + } catch (InterruptedException e) { + Thread.currentThread().interrupt(); + throw new RuntimeException("Interrupted in call to initialize", e); + } + } + + private void initializeCatalogTables() throws InterruptedException, SQLException { + LOG.trace("Creating database tables (if missing) to store iceberg catalog"); + connections.run(conn -> { + DatabaseMetaData dbMeta = conn.getMetaData(); + ResultSet tableExists = dbMeta.getTables(null, null, JdbcUtil.CATALOG_TABLE_NAME, null); + + if (tableExists.next()) { + return true; + } + + LOG.debug("Creating table {} to store iceberg catalog!", JdbcUtil.CATALOG_TABLE_NAME); + return conn.prepareStatement(JdbcUtil.CREATE_CATALOG_TABLE).execute(); + }); + } + + @Override + protected TableOperations newTableOps(TableIdentifier tableIdentifier) { + return new JdbcTableOperations(connections, io, catalogName, tableIdentifier); + } + + @Override + protected String defaultWarehouseLocation(TableIdentifier table) { + return SLASH.join(defaultNamespaceLocation(table.namespace()), table.name()); + } + + @Override + public boolean dropTable(TableIdentifier identifier, boolean purge) { + + int deletedRecords; + try { + deletedRecords = connections.run(conn -> { + try (PreparedStatement sql = conn.prepareStatement(JdbcUtil.DROP_TABLE_SQL)) { + sql.setString(1, catalogName); + sql.setString(2, JdbcUtil.namespaceToString(identifier.namespace())); + sql.setString(3, identifier.name()); + return sql.executeUpdate(); + } + }); + } catch (SQLException e) { + throw new UncheckedSQLException(e, "Failed to drop %s", identifier); + } catch (InterruptedException e) { + Thread.currentThread().interrupt(); + throw new RuntimeException("Interrupted in call to dropTable", e); + } + + if (deletedRecords > 0) { + LOG.info("Successfully dropped table {}.", identifier); Review comment: I'm okay with either debug or info for this one. ########## File path: core/src/main/java/org/apache/iceberg/jdbc/JdbcTableOperations.java ########## @@ -0,0 +1,220 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, + * software distributed under the License is distributed on an + * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY + * KIND, either express or implied. See the License for the + * specific language governing permissions and limitations + * under the License. + */ + +package org.apache.iceberg.jdbc; + +import java.sql.DataTruncation; +import java.sql.PreparedStatement; +import java.sql.ResultSet; +import java.sql.SQLException; +import java.sql.SQLIntegrityConstraintViolationException; +import java.sql.SQLNonTransientConnectionException; +import java.sql.SQLTimeoutException; +import java.sql.SQLTransientConnectionException; +import java.sql.SQLWarning; +import java.util.Map; +import java.util.Objects; +import org.apache.iceberg.BaseMetastoreTableOperations; +import org.apache.iceberg.TableMetadata; +import org.apache.iceberg.catalog.TableIdentifier; +import org.apache.iceberg.exceptions.AlreadyExistsException; +import org.apache.iceberg.exceptions.CommitFailedException; +import org.apache.iceberg.exceptions.NoSuchTableException; +import org.apache.iceberg.io.FileIO; +import org.apache.iceberg.relocated.com.google.common.collect.Maps; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +class JdbcTableOperations extends BaseMetastoreTableOperations { + + private static final Logger LOG = LoggerFactory.getLogger(JdbcTableOperations.class); + private final String catalogName; + private final TableIdentifier tableIdentifier; + private final FileIO fileIO; + private final JdbcClientPool connections; + + protected JdbcTableOperations(JdbcClientPool dbConnPool, FileIO fileIO, String catalogName, + TableIdentifier tableIdentifier) { + this.catalogName = catalogName; + this.tableIdentifier = tableIdentifier; + this.fileIO = fileIO; + this.connections = dbConnPool; + } + + @Override + public void doRefresh() { + Map<String, String> table; + + try { + table = getTable(); + } catch (InterruptedException e) { + Thread.currentThread().interrupt(); + throw new UncheckedInterruptedException(e, "Interrupted during refresh"); + } catch (SQLException e) { + // SQL exception happened when getting table from catalog + throw new UncheckedSQLException(e, "Failed to get table %s from catalog %s", tableIdentifier, catalogName); + } + + // Table not exists AND currentMetadataLocation is not NULL! + if (table.isEmpty() && currentMetadataLocation() != null) { + throw new NoSuchTableException("Failed to get table %s from catalog %s" + + " maybe another process deleted it", tableIdentifier, catalogName); + } + + // Table not exists in the catalog! metadataLocation is null here! + if (table.isEmpty()) { + refreshFromMetadataLocation(null); + return; + } + + // Table exists but metadataLocation is null + if (table.getOrDefault(JdbcUtil.METADATA_LOCATION, null) == null) { + throw new RuntimeException(String.format("Failed to get metadata location of the table %s from catalog %s", + tableIdentifier, catalogName)); + } + + refreshFromMetadataLocation(table.get(JdbcUtil.METADATA_LOCATION)); + } + + @Override + public void doCommit(TableMetadata base, TableMetadata metadata) { + String newMetadataLocation = writeNewMetadata(metadata, currentVersion() + 1); + try { + Map<String, String> table = getTable(); + + if (!table.isEmpty()) { + validateMetadataLocation(table, base); + String oldMetadataLocation = base.metadataFileLocation(); + // Start atomic update + LOG.debug("Committing existing table: {}", tableName()); + updateTable(newMetadataLocation, oldMetadataLocation); + } else { + // table not exists create it + LOG.debug("Committing new table: {}", tableName()); + createTable(newMetadataLocation); + } + + } catch (SQLIntegrityConstraintViolationException e) { + throw new AlreadyExistsException(e, "Table already exists, maybe another process created it"); + } catch (SQLTimeoutException e) { + throw new UncheckedSQLException(e, "Database Connection timeout"); + } catch (SQLTransientConnectionException | SQLNonTransientConnectionException e) { + throw new UncheckedSQLException(e, "Database Connection failed"); + } catch (DataTruncation e) { + throw new UncheckedSQLException(e, "Database data truncation error"); + } catch (SQLWarning e) { + throw new UncheckedSQLException(e, "Database warning"); + } catch (SQLException e) { + throw new UncheckedSQLException(e, "Failed to connect to database"); + } catch (InterruptedException e) { + Thread.currentThread().interrupt(); + throw new UncheckedInterruptedException(e, "Interrupted during commit"); + } + } + + private void updateTable(String newMetadataLocation, String oldMetadataLocation) + throws SQLException, InterruptedException { + int updatedRecords = connections.run(conn -> { + try (PreparedStatement sql = conn.prepareStatement(JdbcUtil.DO_COMMIT_SQL)) { + // UPDATE + sql.setString(1, newMetadataLocation); + sql.setString(2, oldMetadataLocation); + // WHERE + sql.setString(3, catalogName); + sql.setString(4, JdbcUtil.namespaceToString(tableIdentifier.namespace())); + sql.setString(5, tableIdentifier.name()); + sql.setString(6, oldMetadataLocation); + return sql.executeUpdate(); + } + }); + + if (updatedRecords == 1) { + LOG.debug("Successfully committed to existing table: {}", tableIdentifier); + } else { + throw new CommitFailedException("Failed to update the table %s from catalog %s " + + "Maybe another process changed it", tableIdentifier, catalogName); + } + + } + + private void createTable(String newMetadataLocation) throws SQLException, InterruptedException { + int insertRecord = connections.run(conn -> { + try (PreparedStatement sql = conn.prepareStatement(JdbcUtil.DO_COMMIT_CREATE_TABLE_SQL)) { + sql.setString(1, catalogName); + sql.setString(2, JdbcUtil.namespaceToString(tableIdentifier.namespace())); + sql.setString(3, tableIdentifier.name()); + sql.setString(4, newMetadataLocation); + return sql.executeUpdate(); + } + }); + + if (insertRecord == 1) { + LOG.debug("Successfully committed to new table: {}", tableIdentifier); + } else { + throw new CommitFailedException("Failed to create table %s catalog %s", tableIdentifier, catalogName); Review comment: Nit: should be "in catalog" not just "catalog" ########## File path: core/src/main/java/org/apache/iceberg/jdbc/JdbcTableOperations.java ########## @@ -0,0 +1,220 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, + * software distributed under the License is distributed on an + * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY + * KIND, either express or implied. See the License for the + * specific language governing permissions and limitations + * under the License. + */ + +package org.apache.iceberg.jdbc; + +import java.sql.DataTruncation; +import java.sql.PreparedStatement; +import java.sql.ResultSet; +import java.sql.SQLException; +import java.sql.SQLIntegrityConstraintViolationException; +import java.sql.SQLNonTransientConnectionException; +import java.sql.SQLTimeoutException; +import java.sql.SQLTransientConnectionException; +import java.sql.SQLWarning; +import java.util.Map; +import java.util.Objects; +import org.apache.iceberg.BaseMetastoreTableOperations; +import org.apache.iceberg.TableMetadata; +import org.apache.iceberg.catalog.TableIdentifier; +import org.apache.iceberg.exceptions.AlreadyExistsException; +import org.apache.iceberg.exceptions.CommitFailedException; +import org.apache.iceberg.exceptions.NoSuchTableException; +import org.apache.iceberg.io.FileIO; +import org.apache.iceberg.relocated.com.google.common.collect.Maps; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +class JdbcTableOperations extends BaseMetastoreTableOperations { + + private static final Logger LOG = LoggerFactory.getLogger(JdbcTableOperations.class); + private final String catalogName; + private final TableIdentifier tableIdentifier; + private final FileIO fileIO; + private final JdbcClientPool connections; + + protected JdbcTableOperations(JdbcClientPool dbConnPool, FileIO fileIO, String catalogName, + TableIdentifier tableIdentifier) { + this.catalogName = catalogName; + this.tableIdentifier = tableIdentifier; + this.fileIO = fileIO; + this.connections = dbConnPool; + } + + @Override + public void doRefresh() { + Map<String, String> table; + + try { + table = getTable(); + } catch (InterruptedException e) { + Thread.currentThread().interrupt(); + throw new UncheckedInterruptedException(e, "Interrupted during refresh"); + } catch (SQLException e) { + // SQL exception happened when getting table from catalog + throw new UncheckedSQLException(e, "Failed to get table %s from catalog %s", tableIdentifier, catalogName); + } + + // Table not exists AND currentMetadataLocation is not NULL! + if (table.isEmpty() && currentMetadataLocation() != null) { + throw new NoSuchTableException("Failed to get table %s from catalog %s" + + " maybe another process deleted it", tableIdentifier, catalogName); + } + + // Table not exists in the catalog! metadataLocation is null here! + if (table.isEmpty()) { + refreshFromMetadataLocation(null); + return; + } + + // Table exists but metadataLocation is null + if (table.getOrDefault(JdbcUtil.METADATA_LOCATION, null) == null) { + throw new RuntimeException(String.format("Failed to get metadata location of the table %s from catalog %s", + tableIdentifier, catalogName)); + } + + refreshFromMetadataLocation(table.get(JdbcUtil.METADATA_LOCATION)); + } + + @Override + public void doCommit(TableMetadata base, TableMetadata metadata) { + String newMetadataLocation = writeNewMetadata(metadata, currentVersion() + 1); + try { + Map<String, String> table = getTable(); + + if (!table.isEmpty()) { + validateMetadataLocation(table, base); + String oldMetadataLocation = base.metadataFileLocation(); + // Start atomic update + LOG.debug("Committing existing table: {}", tableName()); + updateTable(newMetadataLocation, oldMetadataLocation); + } else { + // table not exists create it + LOG.debug("Committing new table: {}", tableName()); + createTable(newMetadataLocation); + } + + } catch (SQLIntegrityConstraintViolationException e) { + throw new AlreadyExistsException(e, "Table already exists, maybe another process created it"); + } catch (SQLTimeoutException e) { + throw new UncheckedSQLException(e, "Database Connection timeout"); + } catch (SQLTransientConnectionException | SQLNonTransientConnectionException e) { + throw new UncheckedSQLException(e, "Database Connection failed"); + } catch (DataTruncation e) { + throw new UncheckedSQLException(e, "Database data truncation error"); + } catch (SQLWarning e) { + throw new UncheckedSQLException(e, "Database warning"); + } catch (SQLException e) { + throw new UncheckedSQLException(e, "Failed to connect to database"); + } catch (InterruptedException e) { + Thread.currentThread().interrupt(); + throw new UncheckedInterruptedException(e, "Interrupted during commit"); + } + } + + private void updateTable(String newMetadataLocation, String oldMetadataLocation) + throws SQLException, InterruptedException { + int updatedRecords = connections.run(conn -> { + try (PreparedStatement sql = conn.prepareStatement(JdbcUtil.DO_COMMIT_SQL)) { + // UPDATE + sql.setString(1, newMetadataLocation); + sql.setString(2, oldMetadataLocation); + // WHERE + sql.setString(3, catalogName); + sql.setString(4, JdbcUtil.namespaceToString(tableIdentifier.namespace())); + sql.setString(5, tableIdentifier.name()); + sql.setString(6, oldMetadataLocation); + return sql.executeUpdate(); + } + }); + + if (updatedRecords == 1) { + LOG.debug("Successfully committed to existing table: {}", tableIdentifier); + } else { + throw new CommitFailedException("Failed to update the table %s from catalog %s " + + "Maybe another process changed it", tableIdentifier, catalogName); Review comment: Nit: No need for "maybe" here. It is enough to say that the table has changed. ########## File path: core/src/main/java/org/apache/iceberg/jdbc/JdbcTableOperations.java ########## @@ -0,0 +1,220 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, + * software distributed under the License is distributed on an + * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY + * KIND, either express or implied. See the License for the + * specific language governing permissions and limitations + * under the License. + */ + +package org.apache.iceberg.jdbc; + +import java.sql.DataTruncation; +import java.sql.PreparedStatement; +import java.sql.ResultSet; +import java.sql.SQLException; +import java.sql.SQLIntegrityConstraintViolationException; +import java.sql.SQLNonTransientConnectionException; +import java.sql.SQLTimeoutException; +import java.sql.SQLTransientConnectionException; +import java.sql.SQLWarning; +import java.util.Map; +import java.util.Objects; +import org.apache.iceberg.BaseMetastoreTableOperations; +import org.apache.iceberg.TableMetadata; +import org.apache.iceberg.catalog.TableIdentifier; +import org.apache.iceberg.exceptions.AlreadyExistsException; +import org.apache.iceberg.exceptions.CommitFailedException; +import org.apache.iceberg.exceptions.NoSuchTableException; +import org.apache.iceberg.io.FileIO; +import org.apache.iceberg.relocated.com.google.common.collect.Maps; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +class JdbcTableOperations extends BaseMetastoreTableOperations { + + private static final Logger LOG = LoggerFactory.getLogger(JdbcTableOperations.class); + private final String catalogName; + private final TableIdentifier tableIdentifier; + private final FileIO fileIO; + private final JdbcClientPool connections; + + protected JdbcTableOperations(JdbcClientPool dbConnPool, FileIO fileIO, String catalogName, + TableIdentifier tableIdentifier) { + this.catalogName = catalogName; + this.tableIdentifier = tableIdentifier; + this.fileIO = fileIO; + this.connections = dbConnPool; + } + + @Override + public void doRefresh() { + Map<String, String> table; + + try { + table = getTable(); + } catch (InterruptedException e) { + Thread.currentThread().interrupt(); + throw new UncheckedInterruptedException(e, "Interrupted during refresh"); + } catch (SQLException e) { + // SQL exception happened when getting table from catalog + throw new UncheckedSQLException(e, "Failed to get table %s from catalog %s", tableIdentifier, catalogName); + } + + // Table not exists AND currentMetadataLocation is not NULL! + if (table.isEmpty() && currentMetadataLocation() != null) { + throw new NoSuchTableException("Failed to get table %s from catalog %s" + + " maybe another process deleted it", tableIdentifier, catalogName); Review comment: Similar to below, no need for the "maybe" statement. This failed to load a table that it expected to exist. ########## File path: core/src/main/java/org/apache/iceberg/jdbc/JdbcTableOperations.java ########## @@ -0,0 +1,220 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, + * software distributed under the License is distributed on an + * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY + * KIND, either express or implied. See the License for the + * specific language governing permissions and limitations + * under the License. + */ + +package org.apache.iceberg.jdbc; + +import java.sql.DataTruncation; +import java.sql.PreparedStatement; +import java.sql.ResultSet; +import java.sql.SQLException; +import java.sql.SQLIntegrityConstraintViolationException; +import java.sql.SQLNonTransientConnectionException; +import java.sql.SQLTimeoutException; +import java.sql.SQLTransientConnectionException; +import java.sql.SQLWarning; +import java.util.Map; +import java.util.Objects; +import org.apache.iceberg.BaseMetastoreTableOperations; +import org.apache.iceberg.TableMetadata; +import org.apache.iceberg.catalog.TableIdentifier; +import org.apache.iceberg.exceptions.AlreadyExistsException; +import org.apache.iceberg.exceptions.CommitFailedException; +import org.apache.iceberg.exceptions.NoSuchTableException; +import org.apache.iceberg.io.FileIO; +import org.apache.iceberg.relocated.com.google.common.collect.Maps; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +class JdbcTableOperations extends BaseMetastoreTableOperations { + + private static final Logger LOG = LoggerFactory.getLogger(JdbcTableOperations.class); + private final String catalogName; + private final TableIdentifier tableIdentifier; + private final FileIO fileIO; + private final JdbcClientPool connections; + + protected JdbcTableOperations(JdbcClientPool dbConnPool, FileIO fileIO, String catalogName, + TableIdentifier tableIdentifier) { + this.catalogName = catalogName; + this.tableIdentifier = tableIdentifier; + this.fileIO = fileIO; + this.connections = dbConnPool; + } + + @Override + public void doRefresh() { + Map<String, String> table; + + try { + table = getTable(); + } catch (InterruptedException e) { + Thread.currentThread().interrupt(); + throw new UncheckedInterruptedException(e, "Interrupted during refresh"); + } catch (SQLException e) { + // SQL exception happened when getting table from catalog + throw new UncheckedSQLException(e, "Failed to get table %s from catalog %s", tableIdentifier, catalogName); + } + + // Table not exists AND currentMetadataLocation is not NULL! + if (table.isEmpty() && currentMetadataLocation() != null) { + throw new NoSuchTableException("Failed to get table %s from catalog %s" + + " maybe another process deleted it", tableIdentifier, catalogName); + } + + // Table not exists in the catalog! metadataLocation is null here! + if (table.isEmpty()) { + refreshFromMetadataLocation(null); Review comment: Because `currentMetadataLocation()` must be `null` because of the check above, this will do basically nothing. I see that this call is probably here because Hive always calls `refreshFromMetadataLocation`, but I don't think that it is necessary to call it here. I'd probably combine this with the previous check: ```java if (table.isEmpty()) { if (currentMetadataLocation() != null) { throw new NoSuchTableException("Failed to load table %s from catalog %s: dropped by another process", ...); } else { return; } } ``` ########## File path: core/src/main/java/org/apache/iceberg/jdbc/JdbcTableOperations.java ########## @@ -0,0 +1,220 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, + * software distributed under the License is distributed on an + * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY + * KIND, either express or implied. See the License for the + * specific language governing permissions and limitations + * under the License. + */ + +package org.apache.iceberg.jdbc; + +import java.sql.DataTruncation; +import java.sql.PreparedStatement; +import java.sql.ResultSet; +import java.sql.SQLException; +import java.sql.SQLIntegrityConstraintViolationException; +import java.sql.SQLNonTransientConnectionException; +import java.sql.SQLTimeoutException; +import java.sql.SQLTransientConnectionException; +import java.sql.SQLWarning; +import java.util.Map; +import java.util.Objects; +import org.apache.iceberg.BaseMetastoreTableOperations; +import org.apache.iceberg.TableMetadata; +import org.apache.iceberg.catalog.TableIdentifier; +import org.apache.iceberg.exceptions.AlreadyExistsException; +import org.apache.iceberg.exceptions.CommitFailedException; +import org.apache.iceberg.exceptions.NoSuchTableException; +import org.apache.iceberg.io.FileIO; +import org.apache.iceberg.relocated.com.google.common.collect.Maps; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +class JdbcTableOperations extends BaseMetastoreTableOperations { + + private static final Logger LOG = LoggerFactory.getLogger(JdbcTableOperations.class); + private final String catalogName; + private final TableIdentifier tableIdentifier; + private final FileIO fileIO; + private final JdbcClientPool connections; + + protected JdbcTableOperations(JdbcClientPool dbConnPool, FileIO fileIO, String catalogName, + TableIdentifier tableIdentifier) { + this.catalogName = catalogName; + this.tableIdentifier = tableIdentifier; + this.fileIO = fileIO; + this.connections = dbConnPool; + } + + @Override + public void doRefresh() { + Map<String, String> table; + + try { + table = getTable(); + } catch (InterruptedException e) { + Thread.currentThread().interrupt(); + throw new UncheckedInterruptedException(e, "Interrupted during refresh"); + } catch (SQLException e) { + // SQL exception happened when getting table from catalog + throw new UncheckedSQLException(e, "Failed to get table %s from catalog %s", tableIdentifier, catalogName); + } + + // Table not exists AND currentMetadataLocation is not NULL! + if (table.isEmpty() && currentMetadataLocation() != null) { + throw new NoSuchTableException("Failed to get table %s from catalog %s" + + " maybe another process deleted it", tableIdentifier, catalogName); + } + + // Table not exists in the catalog! metadataLocation is null here! + if (table.isEmpty()) { + refreshFromMetadataLocation(null); + return; + } + + // Table exists but metadataLocation is null + if (table.getOrDefault(JdbcUtil.METADATA_LOCATION, null) == null) { Review comment: This `get` is done twice. Can you refactor to use a variable? I think this should also be `IllegalStateException` rather than just a `RuntimeException` so you could do this: ```java String newMetadataLocation = table.get(JdbcUtil.METADATA_LOCATION); Preconditions.checkState(newMetadataLocation != null, "Invalid table %s: metadata location is null", tableIdentifier); refreshFromMetadataLocation(newMetadataLocation); ``` ########## File path: core/src/main/java/org/apache/iceberg/jdbc/JdbcTableOperations.java ########## @@ -0,0 +1,220 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, + * software distributed under the License is distributed on an + * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY + * KIND, either express or implied. See the License for the + * specific language governing permissions and limitations + * under the License. + */ + +package org.apache.iceberg.jdbc; + +import java.sql.DataTruncation; +import java.sql.PreparedStatement; +import java.sql.ResultSet; +import java.sql.SQLException; +import java.sql.SQLIntegrityConstraintViolationException; +import java.sql.SQLNonTransientConnectionException; +import java.sql.SQLTimeoutException; +import java.sql.SQLTransientConnectionException; +import java.sql.SQLWarning; +import java.util.Map; +import java.util.Objects; +import org.apache.iceberg.BaseMetastoreTableOperations; +import org.apache.iceberg.TableMetadata; +import org.apache.iceberg.catalog.TableIdentifier; +import org.apache.iceberg.exceptions.AlreadyExistsException; +import org.apache.iceberg.exceptions.CommitFailedException; +import org.apache.iceberg.exceptions.NoSuchTableException; +import org.apache.iceberg.io.FileIO; +import org.apache.iceberg.relocated.com.google.common.collect.Maps; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +class JdbcTableOperations extends BaseMetastoreTableOperations { + + private static final Logger LOG = LoggerFactory.getLogger(JdbcTableOperations.class); + private final String catalogName; + private final TableIdentifier tableIdentifier; + private final FileIO fileIO; + private final JdbcClientPool connections; + + protected JdbcTableOperations(JdbcClientPool dbConnPool, FileIO fileIO, String catalogName, + TableIdentifier tableIdentifier) { + this.catalogName = catalogName; + this.tableIdentifier = tableIdentifier; + this.fileIO = fileIO; + this.connections = dbConnPool; + } + + @Override + public void doRefresh() { + Map<String, String> table; + + try { + table = getTable(); + } catch (InterruptedException e) { + Thread.currentThread().interrupt(); + throw new UncheckedInterruptedException(e, "Interrupted during refresh"); + } catch (SQLException e) { + // SQL exception happened when getting table from catalog + throw new UncheckedSQLException(e, "Failed to get table %s from catalog %s", tableIdentifier, catalogName); + } + + // Table not exists AND currentMetadataLocation is not NULL! + if (table.isEmpty() && currentMetadataLocation() != null) { + throw new NoSuchTableException("Failed to get table %s from catalog %s" + + " maybe another process deleted it", tableIdentifier, catalogName); + } + + // Table not exists in the catalog! metadataLocation is null here! + if (table.isEmpty()) { + refreshFromMetadataLocation(null); + return; + } + + // Table exists but metadataLocation is null + if (table.getOrDefault(JdbcUtil.METADATA_LOCATION, null) == null) { Review comment: (Also, `getOrDefault` with a `null` default is equivalent to `get`.) ########## File path: core/src/main/java/org/apache/iceberg/jdbc/JdbcTableOperations.java ########## @@ -0,0 +1,220 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, + * software distributed under the License is distributed on an + * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY + * KIND, either express or implied. See the License for the + * specific language governing permissions and limitations + * under the License. + */ + +package org.apache.iceberg.jdbc; + +import java.sql.DataTruncation; +import java.sql.PreparedStatement; +import java.sql.ResultSet; +import java.sql.SQLException; +import java.sql.SQLIntegrityConstraintViolationException; +import java.sql.SQLNonTransientConnectionException; +import java.sql.SQLTimeoutException; +import java.sql.SQLTransientConnectionException; +import java.sql.SQLWarning; +import java.util.Map; +import java.util.Objects; +import org.apache.iceberg.BaseMetastoreTableOperations; +import org.apache.iceberg.TableMetadata; +import org.apache.iceberg.catalog.TableIdentifier; +import org.apache.iceberg.exceptions.AlreadyExistsException; +import org.apache.iceberg.exceptions.CommitFailedException; +import org.apache.iceberg.exceptions.NoSuchTableException; +import org.apache.iceberg.io.FileIO; +import org.apache.iceberg.relocated.com.google.common.collect.Maps; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +class JdbcTableOperations extends BaseMetastoreTableOperations { + + private static final Logger LOG = LoggerFactory.getLogger(JdbcTableOperations.class); + private final String catalogName; + private final TableIdentifier tableIdentifier; + private final FileIO fileIO; + private final JdbcClientPool connections; + + protected JdbcTableOperations(JdbcClientPool dbConnPool, FileIO fileIO, String catalogName, + TableIdentifier tableIdentifier) { + this.catalogName = catalogName; + this.tableIdentifier = tableIdentifier; + this.fileIO = fileIO; + this.connections = dbConnPool; + } + + @Override + public void doRefresh() { + Map<String, String> table; + + try { + table = getTable(); + } catch (InterruptedException e) { + Thread.currentThread().interrupt(); + throw new UncheckedInterruptedException(e, "Interrupted during refresh"); + } catch (SQLException e) { + // SQL exception happened when getting table from catalog + throw new UncheckedSQLException(e, "Failed to get table %s from catalog %s", tableIdentifier, catalogName); + } + + // Table not exists AND currentMetadataLocation is not NULL! + if (table.isEmpty() && currentMetadataLocation() != null) { + throw new NoSuchTableException("Failed to get table %s from catalog %s" + + " maybe another process deleted it", tableIdentifier, catalogName); + } + + // Table not exists in the catalog! metadataLocation is null here! + if (table.isEmpty()) { + refreshFromMetadataLocation(null); + return; + } + + // Table exists but metadataLocation is null + if (table.getOrDefault(JdbcUtil.METADATA_LOCATION, null) == null) { + throw new RuntimeException(String.format("Failed to get metadata location of the table %s from catalog %s", + tableIdentifier, catalogName)); + } + + refreshFromMetadataLocation(table.get(JdbcUtil.METADATA_LOCATION)); + } + + @Override + public void doCommit(TableMetadata base, TableMetadata metadata) { + String newMetadataLocation = writeNewMetadata(metadata, currentVersion() + 1); + try { + Map<String, String> table = getTable(); + + if (!table.isEmpty()) { + validateMetadataLocation(table, base); + String oldMetadataLocation = base.metadataFileLocation(); + // Start atomic update + LOG.debug("Committing existing table: {}", tableName()); + updateTable(newMetadataLocation, oldMetadataLocation); + } else { + // table not exists create it + LOG.debug("Committing new table: {}", tableName()); + createTable(newMetadataLocation); + } + + } catch (SQLIntegrityConstraintViolationException e) { + throw new AlreadyExistsException(e, "Table already exists, maybe another process created it"); + } catch (SQLTimeoutException e) { + throw new UncheckedSQLException(e, "Database Connection timeout"); + } catch (SQLTransientConnectionException | SQLNonTransientConnectionException e) { + throw new UncheckedSQLException(e, "Database Connection failed"); + } catch (DataTruncation e) { + throw new UncheckedSQLException(e, "Database data truncation error"); + } catch (SQLWarning e) { + throw new UncheckedSQLException(e, "Database warning"); + } catch (SQLException e) { + throw new UncheckedSQLException(e, "Failed to connect to database"); + } catch (InterruptedException e) { + Thread.currentThread().interrupt(); + throw new UncheckedInterruptedException(e, "Interrupted during commit"); + } + } + + private void updateTable(String newMetadataLocation, String oldMetadataLocation) + throws SQLException, InterruptedException { + int updatedRecords = connections.run(conn -> { + try (PreparedStatement sql = conn.prepareStatement(JdbcUtil.DO_COMMIT_SQL)) { + // UPDATE + sql.setString(1, newMetadataLocation); + sql.setString(2, oldMetadataLocation); + // WHERE + sql.setString(3, catalogName); + sql.setString(4, JdbcUtil.namespaceToString(tableIdentifier.namespace())); + sql.setString(5, tableIdentifier.name()); + sql.setString(6, oldMetadataLocation); + return sql.executeUpdate(); + } + }); + + if (updatedRecords == 1) { + LOG.debug("Successfully committed to existing table: {}", tableIdentifier); + } else { + throw new CommitFailedException("Failed to update the table %s from catalog %s " + + "Maybe another process changed it", tableIdentifier, catalogName); + } + + } + + private void createTable(String newMetadataLocation) throws SQLException, InterruptedException { + int insertRecord = connections.run(conn -> { + try (PreparedStatement sql = conn.prepareStatement(JdbcUtil.DO_COMMIT_CREATE_TABLE_SQL)) { + sql.setString(1, catalogName); + sql.setString(2, JdbcUtil.namespaceToString(tableIdentifier.namespace())); + sql.setString(3, tableIdentifier.name()); + sql.setString(4, newMetadataLocation); + return sql.executeUpdate(); + } + }); + + if (insertRecord == 1) { + LOG.debug("Successfully committed to new table: {}", tableIdentifier); + } else { + throw new CommitFailedException("Failed to create table %s catalog %s", tableIdentifier, catalogName); + } + } + + private void validateMetadataLocation(Map<String, String> table, TableMetadata base) { + String catalogMetadataLocation = !table.isEmpty() ? table.get(JdbcUtil.METADATA_LOCATION) : null; Review comment: I think this can just be `table.get(JdbcUtil.METADATA_LOCATION)`. ########## File path: core/src/main/java/org/apache/iceberg/jdbc/JdbcTableOperations.java ########## @@ -0,0 +1,220 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, + * software distributed under the License is distributed on an + * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY + * KIND, either express or implied. See the License for the + * specific language governing permissions and limitations + * under the License. + */ + +package org.apache.iceberg.jdbc; + +import java.sql.DataTruncation; +import java.sql.PreparedStatement; +import java.sql.ResultSet; +import java.sql.SQLException; +import java.sql.SQLIntegrityConstraintViolationException; +import java.sql.SQLNonTransientConnectionException; +import java.sql.SQLTimeoutException; +import java.sql.SQLTransientConnectionException; +import java.sql.SQLWarning; +import java.util.Map; +import java.util.Objects; +import org.apache.iceberg.BaseMetastoreTableOperations; +import org.apache.iceberg.TableMetadata; +import org.apache.iceberg.catalog.TableIdentifier; +import org.apache.iceberg.exceptions.AlreadyExistsException; +import org.apache.iceberg.exceptions.CommitFailedException; +import org.apache.iceberg.exceptions.NoSuchTableException; +import org.apache.iceberg.io.FileIO; +import org.apache.iceberg.relocated.com.google.common.collect.Maps; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +class JdbcTableOperations extends BaseMetastoreTableOperations { + + private static final Logger LOG = LoggerFactory.getLogger(JdbcTableOperations.class); + private final String catalogName; + private final TableIdentifier tableIdentifier; + private final FileIO fileIO; + private final JdbcClientPool connections; + + protected JdbcTableOperations(JdbcClientPool dbConnPool, FileIO fileIO, String catalogName, + TableIdentifier tableIdentifier) { + this.catalogName = catalogName; + this.tableIdentifier = tableIdentifier; + this.fileIO = fileIO; + this.connections = dbConnPool; + } + + @Override + public void doRefresh() { + Map<String, String> table; + + try { + table = getTable(); + } catch (InterruptedException e) { + Thread.currentThread().interrupt(); + throw new UncheckedInterruptedException(e, "Interrupted during refresh"); + } catch (SQLException e) { + // SQL exception happened when getting table from catalog + throw new UncheckedSQLException(e, "Failed to get table %s from catalog %s", tableIdentifier, catalogName); + } + + // Table not exists AND currentMetadataLocation is not NULL! + if (table.isEmpty() && currentMetadataLocation() != null) { + throw new NoSuchTableException("Failed to get table %s from catalog %s" + + " maybe another process deleted it", tableIdentifier, catalogName); + } + + // Table not exists in the catalog! metadataLocation is null here! + if (table.isEmpty()) { + refreshFromMetadataLocation(null); + return; + } + + // Table exists but metadataLocation is null + if (table.getOrDefault(JdbcUtil.METADATA_LOCATION, null) == null) { + throw new RuntimeException(String.format("Failed to get metadata location of the table %s from catalog %s", + tableIdentifier, catalogName)); + } + + refreshFromMetadataLocation(table.get(JdbcUtil.METADATA_LOCATION)); + } + + @Override + public void doCommit(TableMetadata base, TableMetadata metadata) { + String newMetadataLocation = writeNewMetadata(metadata, currentVersion() + 1); + try { + Map<String, String> table = getTable(); + + if (!table.isEmpty()) { + validateMetadataLocation(table, base); + String oldMetadataLocation = base.metadataFileLocation(); + // Start atomic update + LOG.debug("Committing existing table: {}", tableName()); + updateTable(newMetadataLocation, oldMetadataLocation); + } else { + // table not exists create it + LOG.debug("Committing new table: {}", tableName()); + createTable(newMetadataLocation); + } + + } catch (SQLIntegrityConstraintViolationException e) { + throw new AlreadyExistsException(e, "Table already exists, maybe another process created it"); + } catch (SQLTimeoutException e) { + throw new UncheckedSQLException(e, "Database Connection timeout"); + } catch (SQLTransientConnectionException | SQLNonTransientConnectionException e) { + throw new UncheckedSQLException(e, "Database Connection failed"); + } catch (DataTruncation e) { + throw new UncheckedSQLException(e, "Database data truncation error"); + } catch (SQLWarning e) { + throw new UncheckedSQLException(e, "Database warning"); + } catch (SQLException e) { + throw new UncheckedSQLException(e, "Failed to connect to database"); + } catch (InterruptedException e) { + Thread.currentThread().interrupt(); + throw new UncheckedInterruptedException(e, "Interrupted during commit"); + } + } + + private void updateTable(String newMetadataLocation, String oldMetadataLocation) + throws SQLException, InterruptedException { + int updatedRecords = connections.run(conn -> { + try (PreparedStatement sql = conn.prepareStatement(JdbcUtil.DO_COMMIT_SQL)) { + // UPDATE + sql.setString(1, newMetadataLocation); + sql.setString(2, oldMetadataLocation); + // WHERE + sql.setString(3, catalogName); + sql.setString(4, JdbcUtil.namespaceToString(tableIdentifier.namespace())); + sql.setString(5, tableIdentifier.name()); + sql.setString(6, oldMetadataLocation); + return sql.executeUpdate(); + } + }); + + if (updatedRecords == 1) { + LOG.debug("Successfully committed to existing table: {}", tableIdentifier); + } else { + throw new CommitFailedException("Failed to update the table %s from catalog %s " + + "Maybe another process changed it", tableIdentifier, catalogName); + } + + } + + private void createTable(String newMetadataLocation) throws SQLException, InterruptedException { + int insertRecord = connections.run(conn -> { + try (PreparedStatement sql = conn.prepareStatement(JdbcUtil.DO_COMMIT_CREATE_TABLE_SQL)) { + sql.setString(1, catalogName); + sql.setString(2, JdbcUtil.namespaceToString(tableIdentifier.namespace())); + sql.setString(3, tableIdentifier.name()); + sql.setString(4, newMetadataLocation); + return sql.executeUpdate(); + } + }); + + if (insertRecord == 1) { + LOG.debug("Successfully committed to new table: {}", tableIdentifier); + } else { + throw new CommitFailedException("Failed to create table %s catalog %s", tableIdentifier, catalogName); + } + } + + private void validateMetadataLocation(Map<String, String> table, TableMetadata base) { + String catalogMetadataLocation = !table.isEmpty() ? table.get(JdbcUtil.METADATA_LOCATION) : null; + String baseMetadataLocation = base != null ? base.metadataFileLocation() : null; + + if (!Objects.equals(baseMetadataLocation, catalogMetadataLocation)) { + throw new CommitFailedException( + "Cannot commit %s because base metadata location '%s' is not same as the current Catalog location '%s'", Review comment: Nit: error message could be more concise: `"Cannot commit %s: metadata location %s has changed from %s"` ########## File path: core/src/main/java/org/apache/iceberg/jdbc/JdbcTableOperations.java ########## @@ -0,0 +1,220 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, + * software distributed under the License is distributed on an + * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY + * KIND, either express or implied. See the License for the + * specific language governing permissions and limitations + * under the License. + */ + +package org.apache.iceberg.jdbc; + +import java.sql.DataTruncation; +import java.sql.PreparedStatement; +import java.sql.ResultSet; +import java.sql.SQLException; +import java.sql.SQLIntegrityConstraintViolationException; +import java.sql.SQLNonTransientConnectionException; +import java.sql.SQLTimeoutException; +import java.sql.SQLTransientConnectionException; +import java.sql.SQLWarning; +import java.util.Map; +import java.util.Objects; +import org.apache.iceberg.BaseMetastoreTableOperations; +import org.apache.iceberg.TableMetadata; +import org.apache.iceberg.catalog.TableIdentifier; +import org.apache.iceberg.exceptions.AlreadyExistsException; +import org.apache.iceberg.exceptions.CommitFailedException; +import org.apache.iceberg.exceptions.NoSuchTableException; +import org.apache.iceberg.io.FileIO; +import org.apache.iceberg.relocated.com.google.common.collect.Maps; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +class JdbcTableOperations extends BaseMetastoreTableOperations { + + private static final Logger LOG = LoggerFactory.getLogger(JdbcTableOperations.class); + private final String catalogName; + private final TableIdentifier tableIdentifier; + private final FileIO fileIO; + private final JdbcClientPool connections; + + protected JdbcTableOperations(JdbcClientPool dbConnPool, FileIO fileIO, String catalogName, + TableIdentifier tableIdentifier) { + this.catalogName = catalogName; + this.tableIdentifier = tableIdentifier; + this.fileIO = fileIO; + this.connections = dbConnPool; + } + + @Override + public void doRefresh() { + Map<String, String> table; + + try { + table = getTable(); + } catch (InterruptedException e) { + Thread.currentThread().interrupt(); + throw new UncheckedInterruptedException(e, "Interrupted during refresh"); + } catch (SQLException e) { + // SQL exception happened when getting table from catalog + throw new UncheckedSQLException(e, "Failed to get table %s from catalog %s", tableIdentifier, catalogName); + } + + // Table not exists AND currentMetadataLocation is not NULL! + if (table.isEmpty() && currentMetadataLocation() != null) { + throw new NoSuchTableException("Failed to get table %s from catalog %s" + + " maybe another process deleted it", tableIdentifier, catalogName); + } + + // Table not exists in the catalog! metadataLocation is null here! + if (table.isEmpty()) { + refreshFromMetadataLocation(null); + return; + } + + // Table exists but metadataLocation is null + if (table.getOrDefault(JdbcUtil.METADATA_LOCATION, null) == null) { + throw new RuntimeException(String.format("Failed to get metadata location of the table %s from catalog %s", + tableIdentifier, catalogName)); + } + + refreshFromMetadataLocation(table.get(JdbcUtil.METADATA_LOCATION)); + } + + @Override + public void doCommit(TableMetadata base, TableMetadata metadata) { + String newMetadataLocation = writeNewMetadata(metadata, currentVersion() + 1); + try { + Map<String, String> table = getTable(); + + if (!table.isEmpty()) { + validateMetadataLocation(table, base); + String oldMetadataLocation = base.metadataFileLocation(); + // Start atomic update + LOG.debug("Committing existing table: {}", tableName()); + updateTable(newMetadataLocation, oldMetadataLocation); + } else { + // table not exists create it + LOG.debug("Committing new table: {}", tableName()); + createTable(newMetadataLocation); + } + + } catch (SQLIntegrityConstraintViolationException e) { + throw new AlreadyExistsException(e, "Table already exists, maybe another process created it"); Review comment: I think this should only be thrown if `currentMetadataLocation()` is `null` because that indicates this is a new table. Otherwise, I think this should throw `UncheckedSQLException`. ########## File path: core/src/main/java/org/apache/iceberg/jdbc/JdbcTableOperations.java ########## @@ -0,0 +1,220 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, + * software distributed under the License is distributed on an + * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY + * KIND, either express or implied. See the License for the + * specific language governing permissions and limitations + * under the License. + */ + +package org.apache.iceberg.jdbc; + +import java.sql.DataTruncation; +import java.sql.PreparedStatement; +import java.sql.ResultSet; +import java.sql.SQLException; +import java.sql.SQLIntegrityConstraintViolationException; +import java.sql.SQLNonTransientConnectionException; +import java.sql.SQLTimeoutException; +import java.sql.SQLTransientConnectionException; +import java.sql.SQLWarning; +import java.util.Map; +import java.util.Objects; +import org.apache.iceberg.BaseMetastoreTableOperations; +import org.apache.iceberg.TableMetadata; +import org.apache.iceberg.catalog.TableIdentifier; +import org.apache.iceberg.exceptions.AlreadyExistsException; +import org.apache.iceberg.exceptions.CommitFailedException; +import org.apache.iceberg.exceptions.NoSuchTableException; +import org.apache.iceberg.io.FileIO; +import org.apache.iceberg.relocated.com.google.common.collect.Maps; +import org.slf4j.Logger; +import org.slf4j.LoggerFactory; + +class JdbcTableOperations extends BaseMetastoreTableOperations { + + private static final Logger LOG = LoggerFactory.getLogger(JdbcTableOperations.class); + private final String catalogName; + private final TableIdentifier tableIdentifier; + private final FileIO fileIO; + private final JdbcClientPool connections; + + protected JdbcTableOperations(JdbcClientPool dbConnPool, FileIO fileIO, String catalogName, + TableIdentifier tableIdentifier) { + this.catalogName = catalogName; + this.tableIdentifier = tableIdentifier; + this.fileIO = fileIO; + this.connections = dbConnPool; + } + + @Override + public void doRefresh() { + Map<String, String> table; + + try { + table = getTable(); + } catch (InterruptedException e) { + Thread.currentThread().interrupt(); + throw new UncheckedInterruptedException(e, "Interrupted during refresh"); + } catch (SQLException e) { + // SQL exception happened when getting table from catalog + throw new UncheckedSQLException(e, "Failed to get table %s from catalog %s", tableIdentifier, catalogName); + } + + // Table not exists AND currentMetadataLocation is not NULL! + if (table.isEmpty() && currentMetadataLocation() != null) { + throw new NoSuchTableException("Failed to get table %s from catalog %s" + + " maybe another process deleted it", tableIdentifier, catalogName); + } + + // Table not exists in the catalog! metadataLocation is null here! + if (table.isEmpty()) { + refreshFromMetadataLocation(null); + return; + } + + // Table exists but metadataLocation is null + if (table.getOrDefault(JdbcUtil.METADATA_LOCATION, null) == null) { + throw new RuntimeException(String.format("Failed to get metadata location of the table %s from catalog %s", + tableIdentifier, catalogName)); + } + + refreshFromMetadataLocation(table.get(JdbcUtil.METADATA_LOCATION)); + } + + @Override + public void doCommit(TableMetadata base, TableMetadata metadata) { + String newMetadataLocation = writeNewMetadata(metadata, currentVersion() + 1); + try { + Map<String, String> table = getTable(); + + if (!table.isEmpty()) { + validateMetadataLocation(table, base); + String oldMetadataLocation = base.metadataFileLocation(); + // Start atomic update + LOG.debug("Committing existing table: {}", tableName()); + updateTable(newMetadataLocation, oldMetadataLocation); + } else { + // table not exists create it + LOG.debug("Committing new table: {}", tableName()); + createTable(newMetadataLocation); + } + + } catch (SQLIntegrityConstraintViolationException e) { + throw new AlreadyExistsException(e, "Table already exists, maybe another process created it"); + } catch (SQLTimeoutException e) { + throw new UncheckedSQLException(e, "Database Connection timeout"); + } catch (SQLTransientConnectionException | SQLNonTransientConnectionException e) { + throw new UncheckedSQLException(e, "Database Connection failed"); + } catch (DataTruncation e) { + throw new UncheckedSQLException(e, "Database data truncation error"); + } catch (SQLWarning e) { + throw new UncheckedSQLException(e, "Database warning"); + } catch (SQLException e) { + throw new UncheckedSQLException(e, "Failed to connect to database"); Review comment: I don't think that "failed to connect" is necessarily correct since the SQL exception is generic. How about "Unknown failure"? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [email protected] --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
