This is an automated email from the ASF dual-hosted git repository.

snuyanzin pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/flink-connector-hive.git

commit 32a4703a47e4802f7b7ffb47e836307814764b5c
Author: Xu Huang <[email protected]>
AuthorDate: Tue Oct 8 10:33:07 2024 +0800

    [hotfix] Remove minor unused deprecated method/class
---
 .../hive/read/HiveBulkFormatAdapter.java           | 48 ----------------------
 .../connectors/hive/read/HiveSourceSplit.java      | 16 --------
 2 files changed, 64 deletions(-)

diff --git 
a/flink-connector-hive/src/main/java/org/apache/flink/connectors/hive/read/HiveBulkFormatAdapter.java
 
b/flink-connector-hive/src/main/java/org/apache/flink/connectors/hive/read/HiveBulkFormatAdapter.java
deleted file mode 100644
index b8dcc00c..00000000
--- 
a/flink-connector-hive/src/main/java/org/apache/flink/connectors/hive/read/HiveBulkFormatAdapter.java
+++ /dev/null
@@ -1,48 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-package org.apache.flink.connectors.hive.read;
-
-import org.apache.flink.connectors.hive.JobConfWrapper;
-import org.apache.flink.table.types.DataType;
-import org.apache.flink.table.types.logical.RowType;
-
-import java.util.List;
-
-/** @deprecated Please use {@link HiveInputFormat}. */
-@Deprecated
-public class HiveBulkFormatAdapter extends HiveInputFormat {
-
-    public HiveBulkFormatAdapter(
-            JobConfWrapper jobConfWrapper,
-            List<String> partitionKeys,
-            String[] fieldNames,
-            DataType[] fieldTypes,
-            String hiveVersion,
-            RowType producedRowType,
-            boolean useMapRedReader) {
-        super(
-                jobConfWrapper,
-                partitionKeys,
-                fieldNames,
-                fieldTypes,
-                hiveVersion,
-                producedRowType,
-                useMapRedReader);
-    }
-}
diff --git 
a/flink-connector-hive/src/main/java/org/apache/flink/connectors/hive/read/HiveSourceSplit.java
 
b/flink-connector-hive/src/main/java/org/apache/flink/connectors/hive/read/HiveSourceSplit.java
index 3e5895bc..7905f949 100644
--- 
a/flink-connector-hive/src/main/java/org/apache/flink/connectors/hive/read/HiveSourceSplit.java
+++ 
b/flink-connector-hive/src/main/java/org/apache/flink/connectors/hive/read/HiveSourceSplit.java
@@ -83,22 +83,6 @@ public class HiveSourceSplit extends FileSourceSplit {
                 checkNotNull(hiveTablePartition, "hiveTablePartition can not 
be null");
     }
 
-    /**
-     * @deprecated You should use {@link #HiveSourceSplit(String, Path, long, 
long, long, long,
-     *     String[], CheckpointedPosition, HiveTablePartition)}
-     */
-    @Deprecated
-    public HiveSourceSplit(
-            String id,
-            Path filePath,
-            long offset,
-            long length,
-            String[] hostnames,
-            @Nullable CheckpointedPosition readerPosition,
-            HiveTablePartition hiveTablePartition) {
-        this(id, filePath, offset, length, 0, 0, hostnames, readerPosition, 
hiveTablePartition);
-    }
-
     public HiveTablePartition getHiveTablePartition() {
         return hiveTablePartition;
     }

Reply via email to