This is an automated email from the ASF dual-hosted git repository. mblow pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/asterixdb.git
commit 2f061939c01b05735a6a51f787a8d2a2f50585d0 Merge: 7156522359 7e7960c090 Author: Michael Blow <[email protected]> AuthorDate: Fri Oct 25 11:51:32 2024 -0400 Merge branch 'gerrit/goldfish' into 'master' Change-Id: I088ea2f4a99970b30cdd15f723f05e2db1da9277 .../optimizerts/queries/query-ASTERIXDB-3512.sqlpp | 44 ++++++++++++++++++++++ .../optimizerts/results/query-ASTERIXDB-3512.plan | 28 ++++++++++++++ .../query-ASTERIXDB-3512.1.ddl.sqlpp | 34 +++++++++++++++++ .../query-ASTERIXDB-3512.2.update.sqlpp | 38 +++++++++++++++++++ .../query-ASTERIXDB-3512.3.query.sqlpp | 38 +++++++++++++++++++ .../query-ASTERIXDB-3512.3.adm | 2 + .../src/test/resources/runtimets/sqlpp_queries.xml | 5 +++ .../runtimets/testsuite_external_dataset_s3.xml | 4 +- .../asterix/common/exceptions/ErrorCode.java | 4 +- .../src/main/resources/asx_errormsg/en.properties | 4 +- .../rules/RemoveUnusedAssignAndAggregateRule.java | 15 ++++++-- 11 files changed, 206 insertions(+), 10 deletions(-) diff --cc asterixdb/asterix-app/src/test/resources/runtimets/testsuite_external_dataset_s3.xml index 949c0a3584,2c7af49eb0..aadc42bc82 --- a/asterixdb/asterix-app/src/test/resources/runtimets/testsuite_external_dataset_s3.xml +++ b/asterixdb/asterix-app/src/test/resources/runtimets/testsuite_external_dataset_s3.xml @@@ -105,29 -80,10 +105,29 @@@ <compilation-unit name="supported-adapter-format-compression"> <output-dir compare="Text">supported-adapter-format-compression</output-dir> <expected-error>ASX1188: Unsupported writing adapter 'AZUREBLOB'. Supported adapters: [gcs, localfs, s3]</expected-error> - <expected-error>ASX1189: Unsupported writing format 'csv'. Supported formats: [json]</expected-error> - <expected-error>ASX1096: Unknown compression scheme rar. Supported schemes are [gzip]</expected-error> + <expected-error>ASX1189: Unsupported writing format 'csv'. Supported formats: [json, parquet]</expected-error> + <expected-error>ASX1202: Unsupported compression scheme rar. Supported schemes for json are [gzip]</expected-error> </compilation-unit> </test-case> + <test-case FilePath="copy-to/negative"> + <compilation-unit name="parquet-error-checks"> + <output-dir compare="Text">parquet-error-checks</output-dir> + <expected-error>ASX1079: Compilation error: TYPE() Expression is required for parquet format</expected-error> + <expected-error>ASX0037: Type mismatch: expected value of type integer, but got the value of type BINARY</expected-error> + <expected-error>HYR0132: Extra field in the result, field 'second' does not exist at 'nested' in the schema</expected-error> + <expected-error>HYR0131: Result does not follow the schema, group type expected but found primitive type at 'nested'</expected-error> + <expected-error>HYR0131: Result does not follow the schema, primitive type expected but found group type at 'name'</expected-error> - <expected-error>ASX1201: Storage units expected for the field 'row-group-size' (e.g., 0.1KB, 100kb, 1mb, 3MB, 8.5GB ...). Provided 'random'</expected-error> - <expected-error>ASX1201: Storage units expected for the field 'page-size' (e.g., 0.1KB, 100kb, 1mb, 3MB, 8.5GB ...). Provided 'random'</expected-error> ++ <expected-error>ASX1206: Storage units expected for the field 'row-group-size' (e.g., 0.1KB, 100kb, 1mb, 3MB, 8.5GB ...). Provided 'random'</expected-error> ++ <expected-error>ASX1206: Storage units expected for the field 'page-size' (e.g., 0.1KB, 100kb, 1mb, 3MB, 8.5GB ...). Provided 'random'</expected-error> + <expected-error>ASX1202: Unsupported compression scheme rar. Supported schemes for parquet are [gzip, snappy, zstd]</expected-error> + <expected-error>ASX1001: Syntax error</expected-error> + <expected-error>ASX1204: 'binary' type not supported in parquet format</expected-error> + <expected-error>ASX1205: Invalid Parquet Writer Version provided '3'. Supported values: [1, 2]</expected-error> + <expected-error>ASX1001: Syntax error</expected-error> + <expected-error>ASX1001: Syntax error</expected-error> + <expected-error>ASX1001: Syntax error</expected-error> + </compilation-unit> + </test-case> <test-case FilePath="copy-to/negative"> <compilation-unit name="empty-over"> <output-dir compare="Text">empty-over</output-dir> diff --cc asterixdb/asterix-common/src/main/java/org/apache/asterix/common/exceptions/ErrorCode.java index 7c89bc91e6,091fb6774f..ea7cc3bae9 --- a/asterixdb/asterix-common/src/main/java/org/apache/asterix/common/exceptions/ErrorCode.java +++ b/asterixdb/asterix-common/src/main/java/org/apache/asterix/common/exceptions/ErrorCode.java @@@ -303,12 -303,8 +303,12 @@@ public enum ErrorCode implements IErro DUPLICATE_FIELD_IN_PRIMARY_KEY(1198), INCOMPATIBLE_FIELDS_IN_PRIMARY_KEY(1199), PREFIX_SHOULD_NOT_START_WITH_SLASH(1200), - ILLEGAL_SIZE_PROVIDED(1201), + INVALID_DELTA_TABLE_FORMAT(1201), - + UNSUPPORTED_WRITER_COMPRESSION_SCHEME(1202), + INVALID_PARQUET_SCHEMA(1203), + TYPE_UNSUPPORTED_PARQUET_WRITE(1204), + INVALID_PARQUET_WRITER_VERSION(1205), - INVALID_DELTA_TABLE_FORMAT(1206), ++ ILLEGAL_SIZE_PROVIDED(1206), // Feed errors DATAFLOW_ILLEGAL_STATE(3001), UTIL_DATAFLOW_UTILS_TUPLE_TOO_LARGE(3002), diff --cc asterixdb/asterix-common/src/main/resources/asx_errormsg/en.properties index d1e01f66f9,8e3ccffff3..3e2f124b27 --- a/asterixdb/asterix-common/src/main/resources/asx_errormsg/en.properties +++ b/asterixdb/asterix-common/src/main/resources/asx_errormsg/en.properties @@@ -305,12 -305,7 +305,12 @@@ 1198 = Duplicate field '%1$s' in primary key 1199 = Fields '%1$s' and '%2$s' are incompatible for primary key 1200 = Prefix should not start with "/". Prefix: '%1$s' - 1201 = Storage units expected for the field '%1$s' (e.g., 0.1KB, 100kb, 1mb, 3MB, 8.5GB ...). Provided '%2$s' + 1201 = Supported file format for 'delta' tables is 'parquet', but '%1$s' was provided. +1202 = Unsupported compression scheme %1$s. Supported schemes for %2$s are %3$s +1203 = Invalid schema provided: '%1$s' +1204 = '%1$s' type not supported in parquet format +1205 = Invalid Parquet Writer Version provided '%1$s'. Supported values: %2$s - 1206 = Supported file format for 'delta' tables is 'parquet', but '%1$s' was provided. ++1206 = Storage units expected for the field '%1$s' (e.g., 0.1KB, 100kb, 1mb, 3MB, 8.5GB ...). Provided '%2$s' # Feed Errors 3001 = Illegal state.
