This is an automated email from the ASF dual-hosted git repository. vinish pushed a commit to branch fix-artifact in repository https://gitbox.apache.org/repos/asf/incubator-xtable.git
commit 9fe6add3a09efaf7d28cf0d6a1acde0296950dc8 Author: Vinish Reddy <[email protected]> AuthorDate: Wed Aug 7 23:19:51 2024 +0530 Revert artifact name with incubator prefix --- README.md | 2 +- demo/notebook/demo.ipynb | 6 +++--- demo/start_demo.sh | 6 +++--- pom.xml | 8 ++++---- website/docs/biglake-metastore.md | 4 ++-- website/docs/bigquery.md | 4 ++-- website/docs/fabric.md | 2 +- website/docs/glue-catalog.md | 4 ++-- website/docs/hms.md | 4 ++-- website/docs/how-to.md | 4 ++-- website/docs/unity-catalog.md | 4 ++-- xtable-api/pom.xml | 4 ++-- xtable-core/pom.xml | 8 ++++---- xtable-hudi-support/pom.xml | 4 ++-- xtable-hudi-support/xtable-hudi-support-extensions/README.md | 6 +++--- xtable-hudi-support/xtable-hudi-support-extensions/pom.xml | 8 ++++---- xtable-hudi-support/xtable-hudi-support-utils/pom.xml | 4 ++-- xtable-utilities/pom.xml | 8 ++++---- 18 files changed, 45 insertions(+), 45 deletions(-) diff --git a/README.md b/README.md index df8bfc35..a150af49 100644 --- a/README.md +++ b/README.md @@ -110,7 +110,7 @@ catalogOptions: # all other options are passed through in a map key1: value1 key2: value2 ``` -5. run with `java -jar incubator-xtable-utilities/target/xtable-utilities-0.1.0-SNAPSHOT-bundled.jar --datasetConfig my_config.yaml [--hadoopConfig hdfs-site.xml] [--convertersConfig converters.yaml] [--icebergCatalogConfig catalog.yaml]` +5. run with `java -jar xtable-utilities/target/xtable-utilities-0.1.0-SNAPSHOT-bundled.jar --datasetConfig my_config.yaml [--hadoopConfig hdfs-site.xml] [--convertersConfig converters.yaml] [--icebergCatalogConfig catalog.yaml]` The bundled jar includes hadoop dependencies for AWS, Azure, and GCP. Sample hadoop configurations for configuring the converters can be found in the [xtable-hadoop-defaults.xml](https://github.com/apache/incubator-xtable/blob/main/utilities/src/main/resources/xtable-hadoop-defaults.xml) file. The custom hadoop configurations can be passed in with the `--hadoopConfig [custom-hadoop-config-file]` option. diff --git a/demo/notebook/demo.ipynb b/demo/notebook/demo.ipynb index c49714c1..40aa55ea 100644 --- a/demo/notebook/demo.ipynb +++ b/demo/notebook/demo.ipynb @@ -27,9 +27,9 @@ "import $ivy.`org.apache.hudi:hudi-spark3.2-bundle_2.12:0.14.0`\n", "import $ivy.`org.apache.hudi:hudi-java-client:0.14.0`\n", "import $ivy.`io.delta:delta-core_2.12:2.0.2`\n", - "import $cp.`/home/jars/incubator-xtable-core-0.1.0-SNAPSHOT.jar`\n", - "import $cp.`/home/jars/incubator-xtable-api-0.1.0-SNAPSHOT.jar`\n", - "import $cp.`/home/jars/incubator-xtable-hudi-support-utils-0.1.0-SNAPSHOT.jar`\n", + "import $cp.`/home/jars/xtable-core-0.1.0-SNAPSHOT.jar`\n", + "import $cp.`/home/jars/xtable-api-0.1.0-SNAPSHOT.jar`\n", + "import $cp.`/home/jars/xtable-hudi-support-utils-0.1.0-SNAPSHOT.jar`\n", "import $ivy.`org.apache.iceberg:iceberg-hive-runtime:1.3.1`\n", "import $ivy.`io.trino:trino-jdbc:431`\n", "import java.util._\n", diff --git a/demo/start_demo.sh b/demo/start_demo.sh index 13e799c9..2232ae85 100755 --- a/demo/start_demo.sh +++ b/demo/start_demo.sh @@ -23,9 +23,9 @@ cd $XTABLE_HOME mvn install -am -pl xtable-core -DskipTests -T 2 mkdir -p demo/jars -cp xtable-hudi-support/xtable-hudi-support-utils/target/incubator-xtable-hudi-support-utils-0.1.0-SNAPSHOT.jar demo/jars -cp xtable-api/target/incubator-xtable-api-0.1.0-SNAPSHOT.jar demo/jars -cp xtable-core/target/incubator-xtable-core-0.1.0-SNAPSHOT.jar demo/jars +cp xtable-hudi-support/xtable-hudi-support-utils/target/xtable-hudi-support-utils-0.1.0-SNAPSHOT.jar demo/jars +cp xtable-api/target/xtable-api-0.1.0-SNAPSHOT.jar demo/jars +cp xtable-core/target/xtable-core-0.1.0-SNAPSHOT.jar demo/jars cd demo docker-compose up diff --git a/pom.xml b/pom.xml index 8e864fc3..b3b5572f 100644 --- a/pom.xml +++ b/pom.xml @@ -21,7 +21,7 @@ <modelVersion>4.0.0</modelVersion> <groupId>org.apache.xtable</groupId> - <artifactId>incubator-xtable</artifactId> + <artifactId>xtable</artifactId> <name>xtable</name> <parent> @@ -88,17 +88,17 @@ <dependencies> <dependency> <groupId>org.apache.xtable</groupId> - <artifactId>incubator-xtable-api</artifactId> + <artifactId>xtable-api</artifactId> <version>${project.version}</version> </dependency> <dependency> <groupId>org.apache.xtable</groupId> - <artifactId>incubator-xtable-core</artifactId> + <artifactId>xtable-core</artifactId> <version>${project.version}</version> </dependency> <dependency> <groupId>org.apache.xtable</groupId> - <artifactId>incubator-xtable-hudi-support-utils</artifactId> + <artifactId>xtable-hudi-support-utils</artifactId> <version>${project.version}</version> </dependency> diff --git a/website/docs/biglake-metastore.md b/website/docs/biglake-metastore.md index 629afb9a..db10daa2 100644 --- a/website/docs/biglake-metastore.md +++ b/website/docs/biglake-metastore.md @@ -25,7 +25,7 @@ This document walks through the steps to register an Apache XTable™ (Incubatin export GOOGLE_APPLICATION_CREDENTIALS=/path/to/service_account_key.json ``` 5. Clone the Apache XTable™ (Incubating) [repository](https://github.com/apache/incubator-xtable) and create the - `incubator-xtable-utilities-0.1.0-SNAPSHOT-bundled.jar` by following the steps on the [Installation page](/docs/setup) + `xtable-utilities-0.1.0-SNAPSHOT-bundled.jar` by following the steps on the [Installation page](/docs/setup) 6. Download the [BigLake Iceberg JAR](gs://spark-lib/biglake/biglake-catalog-iceberg1.2.0-0.1.0-with-dependencies.jar) locally. Apache XTable™ (Incubating) requires the JAR to be present in the classpath. @@ -117,7 +117,7 @@ catalogOptions: From your terminal under the cloned Apache XTable™ (Incubating) directory, run the sync process using the below command. ```shell md title="shell" -java -cp xtable-utilities/target/incubator-xtable-utilities-0.1.0-SNAPSHOT-bundled.jar:/path/to/downloaded/biglake-catalog-iceberg1.2.0-0.1.0-with-dependencies.jar org.apache.xtable.utilities.RunSync --datasetConfig my_config.yaml --icebergCatalogConfig catalog.yaml +java -cp xtable-utilities/target/xtable-utilities-0.1.0-SNAPSHOT-bundled.jar:/path/to/downloaded/biglake-catalog-iceberg1.2.0-0.1.0-with-dependencies.jar org.apache.xtable.utilities.RunSync --datasetConfig my_config.yaml --icebergCatalogConfig catalog.yaml ``` :::tip Note: diff --git a/website/docs/bigquery.md b/website/docs/bigquery.md index b7c6b13e..46a928aa 100644 --- a/website/docs/bigquery.md +++ b/website/docs/bigquery.md @@ -35,9 +35,9 @@ If you are not planning on using Iceberg, then you do not need to add these to y ::: #### Steps to add additional configurations to the Hudi writers: -1. Add the extensions jar (`incubator-xtable-hudi-extensions-0.1.0-SNAPSHOT-bundled.jar`) to your class path +1. Add the extensions jar (`xtable-hudi-extensions-0.1.0-SNAPSHOT-bundled.jar`) to your class path For example, if you're using the Hudi [quick-start guide](https://hudi.apache.org/docs/quick-start-guide#spark-shellsql) - for spark you can just add `--jars incubator-xtable-hudi-extensions-0.1.0-SNAPSHOT-bundled.jar` to the end of the command. + for spark you can just add `--jars xtable-hudi-extensions-0.1.0-SNAPSHOT-bundled.jar` to the end of the command. 2. Set the following configurations in your writer options: ```shell md title="shell" hoodie.avro.write.support.class: org.apache.xtable.hudi.extensions.HoodieAvroWriteSupportWithFieldIds diff --git a/website/docs/fabric.md b/website/docs/fabric.md index 6f1f44f1..7bea88ee 100644 --- a/website/docs/fabric.md +++ b/website/docs/fabric.md @@ -98,7 +98,7 @@ An example hadoop configuration for authenticating to ADLS storage account is as ``` ```shell md title="shell" -java -jar xtable-utilities/target/incubator-xtable-utilities-0.1.0-SNAPSHOT-bundled.jar --datasetConfig my_config.yaml --hadoopConfig hadoop.xml +java -jar xtable-utilities/target/xtable-utilities-0.1.0-SNAPSHOT-bundled.jar --datasetConfig my_config.yaml --hadoopConfig hadoop.xml ``` Running the above command will translate the table `people` in Iceberg or Hudi format to Delta Lake format. To validate diff --git a/website/docs/glue-catalog.md b/website/docs/glue-catalog.md index 99ac4da0..62468ed9 100644 --- a/website/docs/glue-catalog.md +++ b/website/docs/glue-catalog.md @@ -19,7 +19,7 @@ This document walks through the steps to register an Apache XTable™ (Incubatin also set up access credentials by following the steps [here](https://docs.aws.amazon.com/cli/latest/userguide/getting-started-quickstart.html) 3. Clone the Apache XTable™ (Incubating) [repository](https://github.com/apache/incubator-xtable) and create the - `incubator-xtable-utilities-0.1.0-SNAPSHOT-bundled.jar` by following the steps on the [Installation page](/docs/setup) + `xtable-utilities-0.1.0-SNAPSHOT-bundled.jar` by following the steps on the [Installation page](/docs/setup) ## Steps ### Running sync @@ -84,7 +84,7 @@ Replace with appropriate values for `sourceFormat`, `tableBasePath` and `tableNa From your terminal under the cloned xtable directory, run the sync process using the below command. ```shell md title="shell" - java -jar xtable-utilities/target/incubator-xtable-utilities-0.1.0-SNAPSHOT-bundled.jar --datasetConfig my_config.yaml + java -jar xtable-utilities/target/xtable-utilities-0.1.0-SNAPSHOT-bundled.jar --datasetConfig my_config.yaml ``` :::tip Note: diff --git a/website/docs/hms.md b/website/docs/hms.md index 5574de43..2698c839 100644 --- a/website/docs/hms.md +++ b/website/docs/hms.md @@ -17,7 +17,7 @@ This document walks through the steps to register an Apache XTable™ (Incubatin or a distributed system like Amazon EMR, Google Cloud's Dataproc, Azure HDInsight etc. This is a required step to register the table in HMS using a Spark client. 3. Clone the XTable™ (Incubating) [repository](https://github.com/apache/incubator-xtable) and create the - `incubator-xtable-utilities-0.1.0-SNAPSHOT-bundled.jar` by following the steps on the [Installation page](/docs/setup) + `xtable-utilities-0.1.0-SNAPSHOT-bundled.jar` by following the steps on the [Installation page](/docs/setup) 4. This guide also assumes that you have configured the Hive Metastore locally or on EMR/Dataproc/HDInsight and is already running. @@ -88,7 +88,7 @@ datasets: From your terminal under the cloned Apache XTable™ (Incubating) directory, run the sync process using the below command. ```shell md title="shell" -java -jar xtable-utilities/target/incubator-xtable-utilities-0.1.0-SNAPSHOT-bundled.jar --datasetConfig my_config.yaml +java -jar xtable-utilities/target/xtable-utilities-0.1.0-SNAPSHOT-bundled.jar --datasetConfig my_config.yaml ``` :::tip Note: diff --git a/website/docs/how-to.md b/website/docs/how-to.md index e3b19de6..5e457c0f 100644 --- a/website/docs/how-to.md +++ b/website/docs/how-to.md @@ -24,7 +24,7 @@ history to enable proper point in time queries. 1. A compute instance where you can run Apache Spark. This can be your local machine, docker, or a distributed service like Amazon EMR, Google Cloud's Dataproc, Azure HDInsight etc 2. Clone the Apache XTable™ (Incubating) [repository](https://github.com/apache/incubator-xtable) and create the - `incubator-xtable-utilities-0.1.0-SNAPSHOT-bundled.jar` by following the steps on the [Installation page](/docs/setup) + `xtable-utilities-0.1.0-SNAPSHOT-bundled.jar` by following the steps on the [Installation page](/docs/setup) 3. Optional: Setup access to write to and/or read from distributed storage services like: * Amazon S3 by following the steps [here](https://docs.aws.amazon.com/cli/latest/userguide/getting-started-install.html) to install AWSCLIv2 @@ -351,7 +351,7 @@ Authentication for GCP requires service account credentials to be exported. i.e. In your terminal under the cloned Apache XTable™ (Incubating) directory, run the below command. ```shell md title="shell" -java -jar xtable-utilities/target/incubator-xtable-utilities-0.1.0-SNAPSHOT-bundled.jar --datasetConfig my_config.yaml +java -jar xtable-utilities/target/xtable-utilities-0.1.0-SNAPSHOT-bundled.jar --datasetConfig my_config.yaml ``` **Optional:** diff --git a/website/docs/unity-catalog.md b/website/docs/unity-catalog.md index aeb338d3..2467a321 100644 --- a/website/docs/unity-catalog.md +++ b/website/docs/unity-catalog.md @@ -17,7 +17,7 @@ This document walks through the steps to register an Apache XTable™ (Incubatin 3. Create a Unity Catalog metastore in Databricks as outlined [here](https://docs.gcp.databricks.com/data-governance/unity-catalog/create-metastore.html#create-a-unity-catalog-metastore). 4. Create an external location in Databricks as outlined [here](https://docs.databricks.com/en/sql/language-manual/sql-ref-syntax-ddl-create-location.html). 5. Clone the Apache XTable™ (Incubating) [repository](https://github.com/apache/incubator-xtable) and create the - `incubator-xtable-utilities-0.1.0-SNAPSHOT-bundled.jar` by following the steps on the [Installation page](/docs/setup) + `xtable-utilities-0.1.0-SNAPSHOT-bundled.jar` by following the steps on the [Installation page](/docs/setup) ## Pre-requisites (for open-source Unity Catalog) 1. Source table(s) (Hudi/Iceberg) already written to external storage locations like S3/GCS/ADLS or local. @@ -48,7 +48,7 @@ datasets: From your terminal under the cloned Apache XTable™ (Incubating) directory, run the sync process using the below command. ```shell md title="shell" -java -jar xtable-utilities/target/incubator-xtable-utilities-0.1.0-SNAPSHOT-bundled.jar --datasetConfig my_config.yaml +java -jar xtable-utilities/target/xtable-utilities-0.1.0-SNAPSHOT-bundled.jar --datasetConfig my_config.yaml ``` :::tip Note: diff --git a/xtable-api/pom.xml b/xtable-api/pom.xml index 5de330c8..71306aee 100644 --- a/xtable-api/pom.xml +++ b/xtable-api/pom.xml @@ -19,12 +19,12 @@ xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd"> <modelVersion>4.0.0</modelVersion> - <artifactId>incubator-xtable-api</artifactId> + <artifactId>xtable-api</artifactId> <name>xtable-api</name> <parent> <groupId>org.apache.xtable</groupId> - <artifactId>incubator-xtable</artifactId> + <artifactId>xtable</artifactId> <version>0.1.0-SNAPSHOT</version> </parent> diff --git a/xtable-core/pom.xml b/xtable-core/pom.xml index cd4fa2ee..f505d265 100644 --- a/xtable-core/pom.xml +++ b/xtable-core/pom.xml @@ -19,23 +19,23 @@ xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd"> <modelVersion>4.0.0</modelVersion> - <artifactId>incubator-xtable-core</artifactId> + <artifactId>xtable-core</artifactId> <name>xtable-core</name> <parent> <groupId>org.apache.xtable</groupId> - <artifactId>incubator-xtable</artifactId> + <artifactId>xtable</artifactId> <version>0.1.0-SNAPSHOT</version> </parent> <dependencies> <dependency> <groupId>org.apache.xtable</groupId> - <artifactId>incubator-xtable-api</artifactId> + <artifactId>xtable-api</artifactId> </dependency> <dependency> <groupId>org.apache.xtable</groupId> - <artifactId>incubator-xtable-hudi-support-utils</artifactId> + <artifactId>xtable-hudi-support-utils</artifactId> </dependency> <dependency> <groupId>com.fasterxml.jackson.core</groupId> diff --git a/xtable-hudi-support/pom.xml b/xtable-hudi-support/pom.xml index 16daa629..9dd29d3c 100644 --- a/xtable-hudi-support/pom.xml +++ b/xtable-hudi-support/pom.xml @@ -21,11 +21,11 @@ <modelVersion>4.0.0</modelVersion> <parent> <groupId>org.apache.xtable</groupId> - <artifactId>incubator-xtable</artifactId> + <artifactId>xtable</artifactId> <version>0.1.0-SNAPSHOT</version> </parent> - <artifactId>incubator-xtable-hudi-support</artifactId> + <artifactId>xtable-hudi-support</artifactId> <packaging>pom</packaging> diff --git a/xtable-hudi-support/xtable-hudi-support-extensions/README.md b/xtable-hudi-support/xtable-hudi-support-extensions/README.md index 42079c54..6e4e3575 100644 --- a/xtable-hudi-support/xtable-hudi-support-extensions/README.md +++ b/xtable-hudi-support/xtable-hudi-support-extensions/README.md @@ -21,8 +21,8 @@ ### When should you use them? The Hudi extensions provide the ability to add field IDs to the parquet schema when writing with Hudi. This is a requirement for some engines, like BigQuery and Snowflake, when reading an Iceberg table. If you are not planning on using Iceberg, then you do not need to add these to your Hudi writers. ### How do you use them? -1. Add the extensions jar (`incubator-xtable-hudi-extensions-0.1.0-SNAPSHOT-bundled.jar`) to your class path. -For example, if you're using the Hudi [quick-start guide](https://hudi.apache.org/docs/quick-start-guide#spark-shellsql) for spark you can just add `--jars incubator-xtable-hudi-extensions-0.1.0-SNAPSHOT-bundled.jar` to the end of the command. +1. Add the extensions jar (`xtable-hudi-extensions-0.1.0-SNAPSHOT-bundled.jar`) to your class path. +For example, if you're using the Hudi [quick-start guide](https://hudi.apache.org/docs/quick-start-guide#spark-shellsql) for spark you can just add `--jars xtable-hudi-extensions-0.1.0-SNAPSHOT-bundled.jar` to the end of the command. 2. Set the following configurations in your writer options: `hoodie.avro.write.support.class: org.apache.xtable.hudi.extensions.HoodieAvroWriteSupportWithFieldIds` `hoodie.client.init.callback.classes: org.apache.xtable.hudi.extensions.AddFieldIdsClientInitCallback` @@ -33,7 +33,7 @@ For example, if you're using the Hudi [quick-start guide](https://hudi.apache.or ### When should you use them? If you want to use XTable with Hudi [streaming ingestion](https://hudi.apache.org/docs/hoodie_streaming_ingestion) to sync each commit into other table formats. ### How do you use them? -1. Add the extensions jar (`incubator-xtable-hudi-extensions-0.1.0-SNAPSHOT-bundled.jar`) to your class path. +1. Add the extensions jar (`xtable-hudi-extensions-0.1.0-SNAPSHOT-bundled.jar`) to your class path. 2. Add `org.apache.xtable.hudi.sync.XTableSyncTool` to your list of sync classes 3. Set the following configurations based on your preferences: `hoodie.xtable.formats.to.sync: "ICEBERG,DELTA"` (or simply use one format) diff --git a/xtable-hudi-support/xtable-hudi-support-extensions/pom.xml b/xtable-hudi-support/xtable-hudi-support-extensions/pom.xml index e8b4a6a8..f149cb62 100644 --- a/xtable-hudi-support/xtable-hudi-support-extensions/pom.xml +++ b/xtable-hudi-support/xtable-hudi-support-extensions/pom.xml @@ -19,22 +19,22 @@ xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd"> <modelVersion>4.0.0</modelVersion> - <artifactId>incubator-xtable-hudi-support-extensions</artifactId> + <artifactId>xtable-hudi-support-extensions</artifactId> <parent> <groupId>org.apache.xtable</groupId> - <artifactId>incubator-xtable-hudi-support</artifactId> + <artifactId>xtable-hudi-support</artifactId> <version>0.1.0-SNAPSHOT</version> </parent> <dependencies> <dependency> <groupId>org.apache.xtable</groupId> - <artifactId>incubator-xtable-hudi-support-utils</artifactId> + <artifactId>xtable-hudi-support-utils</artifactId> </dependency> <dependency> <groupId>org.apache.xtable</groupId> - <artifactId>incubator-xtable-core</artifactId> + <artifactId>xtable-core</artifactId> </dependency> <!-- Logging API --> diff --git a/xtable-hudi-support/xtable-hudi-support-utils/pom.xml b/xtable-hudi-support/xtable-hudi-support-utils/pom.xml index e0f38430..5a44bde8 100644 --- a/xtable-hudi-support/xtable-hudi-support-utils/pom.xml +++ b/xtable-hudi-support/xtable-hudi-support-utils/pom.xml @@ -19,11 +19,11 @@ xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd"> <modelVersion>4.0.0</modelVersion> - <artifactId>incubator-xtable-hudi-support-utils</artifactId> + <artifactId>xtable-hudi-support-utils</artifactId> <parent> <groupId>org.apache.xtable</groupId> - <artifactId>incubator-xtable-hudi-support</artifactId> + <artifactId>xtable-hudi-support</artifactId> <version>0.1.0-SNAPSHOT</version> </parent> diff --git a/xtable-utilities/pom.xml b/xtable-utilities/pom.xml index 61b1dfed..9492aea3 100644 --- a/xtable-utilities/pom.xml +++ b/xtable-utilities/pom.xml @@ -20,21 +20,21 @@ xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd"> <parent> <groupId>org.apache.xtable</groupId> - <artifactId>incubator-xtable</artifactId> + <artifactId>xtable</artifactId> <version>0.1.0-SNAPSHOT</version> </parent> <modelVersion>4.0.0</modelVersion> - <artifactId>incubator-xtable-utilities</artifactId> + <artifactId>xtable-utilities</artifactId> <dependencies> <dependency> <groupId>org.apache.xtable</groupId> - <artifactId>incubator-xtable-api</artifactId> + <artifactId>xtable-api</artifactId> </dependency> <dependency> <groupId>org.apache.xtable</groupId> - <artifactId>incubator-xtable-core</artifactId> + <artifactId>xtable-core</artifactId> </dependency> <!-- command line arg parsing -->
