This is an automated email from the ASF dual-hosted git repository. snuyanzin pushed a commit to branch main in repository https://gitbox.apache.org/repos/asf/flink-connector-hive.git
commit 32a4703a47e4802f7b7ffb47e836307814764b5c Author: Xu Huang <[email protected]> AuthorDate: Tue Oct 8 10:33:07 2024 +0800 [hotfix] Remove minor unused deprecated method/class --- .../hive/read/HiveBulkFormatAdapter.java | 48 ---------------------- .../connectors/hive/read/HiveSourceSplit.java | 16 -------- 2 files changed, 64 deletions(-) diff --git a/flink-connector-hive/src/main/java/org/apache/flink/connectors/hive/read/HiveBulkFormatAdapter.java b/flink-connector-hive/src/main/java/org/apache/flink/connectors/hive/read/HiveBulkFormatAdapter.java deleted file mode 100644 index b8dcc00c..00000000 --- a/flink-connector-hive/src/main/java/org/apache/flink/connectors/hive/read/HiveBulkFormatAdapter.java +++ /dev/null @@ -1,48 +0,0 @@ -/* - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -package org.apache.flink.connectors.hive.read; - -import org.apache.flink.connectors.hive.JobConfWrapper; -import org.apache.flink.table.types.DataType; -import org.apache.flink.table.types.logical.RowType; - -import java.util.List; - -/** @deprecated Please use {@link HiveInputFormat}. */ -@Deprecated -public class HiveBulkFormatAdapter extends HiveInputFormat { - - public HiveBulkFormatAdapter( - JobConfWrapper jobConfWrapper, - List<String> partitionKeys, - String[] fieldNames, - DataType[] fieldTypes, - String hiveVersion, - RowType producedRowType, - boolean useMapRedReader) { - super( - jobConfWrapper, - partitionKeys, - fieldNames, - fieldTypes, - hiveVersion, - producedRowType, - useMapRedReader); - } -} diff --git a/flink-connector-hive/src/main/java/org/apache/flink/connectors/hive/read/HiveSourceSplit.java b/flink-connector-hive/src/main/java/org/apache/flink/connectors/hive/read/HiveSourceSplit.java index 3e5895bc..7905f949 100644 --- a/flink-connector-hive/src/main/java/org/apache/flink/connectors/hive/read/HiveSourceSplit.java +++ b/flink-connector-hive/src/main/java/org/apache/flink/connectors/hive/read/HiveSourceSplit.java @@ -83,22 +83,6 @@ public class HiveSourceSplit extends FileSourceSplit { checkNotNull(hiveTablePartition, "hiveTablePartition can not be null"); } - /** - * @deprecated You should use {@link #HiveSourceSplit(String, Path, long, long, long, long, - * String[], CheckpointedPosition, HiveTablePartition)} - */ - @Deprecated - public HiveSourceSplit( - String id, - Path filePath, - long offset, - long length, - String[] hostnames, - @Nullable CheckpointedPosition readerPosition, - HiveTablePartition hiveTablePartition) { - this(id, filePath, offset, length, 0, 0, hostnames, readerPosition, hiveTablePartition); - } - public HiveTablePartition getHiveTablePartition() { return hiveTablePartition; }
