This is an automated email from the ASF dual-hosted git repository.

mblow pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/asterixdb.git

commit 2f061939c01b05735a6a51f787a8d2a2f50585d0
Merge: 7156522359 7e7960c090
Author: Michael Blow <[email protected]>
AuthorDate: Fri Oct 25 11:51:32 2024 -0400

    Merge branch 'gerrit/goldfish' into 'master'
    
    Change-Id: I088ea2f4a99970b30cdd15f723f05e2db1da9277

 .../optimizerts/queries/query-ASTERIXDB-3512.sqlpp | 44 ++++++++++++++++++++++
 .../optimizerts/results/query-ASTERIXDB-3512.plan  | 28 ++++++++++++++
 .../query-ASTERIXDB-3512.1.ddl.sqlpp               | 34 +++++++++++++++++
 .../query-ASTERIXDB-3512.2.update.sqlpp            | 38 +++++++++++++++++++
 .../query-ASTERIXDB-3512.3.query.sqlpp             | 38 +++++++++++++++++++
 .../query-ASTERIXDB-3512.3.adm                     |  2 +
 .../src/test/resources/runtimets/sqlpp_queries.xml |  5 +++
 .../runtimets/testsuite_external_dataset_s3.xml    |  4 +-
 .../asterix/common/exceptions/ErrorCode.java       |  4 +-
 .../src/main/resources/asx_errormsg/en.properties  |  4 +-
 .../rules/RemoveUnusedAssignAndAggregateRule.java  | 15 ++++++--
 11 files changed, 206 insertions(+), 10 deletions(-)

diff --cc 
asterixdb/asterix-app/src/test/resources/runtimets/testsuite_external_dataset_s3.xml
index 949c0a3584,2c7af49eb0..aadc42bc82
--- 
a/asterixdb/asterix-app/src/test/resources/runtimets/testsuite_external_dataset_s3.xml
+++ 
b/asterixdb/asterix-app/src/test/resources/runtimets/testsuite_external_dataset_s3.xml
@@@ -105,29 -80,10 +105,29 @@@
        <compilation-unit name="supported-adapter-format-compression">
          <output-dir 
compare="Text">supported-adapter-format-compression</output-dir>
          <expected-error>ASX1188: Unsupported writing adapter 'AZUREBLOB'. 
Supported adapters: [gcs, localfs, s3]</expected-error>
 -        <expected-error>ASX1189: Unsupported writing format 'csv'. Supported 
formats: [json]</expected-error>
 -        <expected-error>ASX1096: Unknown compression scheme rar. Supported 
schemes are [gzip]</expected-error>
 +        <expected-error>ASX1189: Unsupported writing format 'csv'. Supported 
formats: [json, parquet]</expected-error>
 +        <expected-error>ASX1202: Unsupported compression scheme rar. 
Supported schemes for json are [gzip]</expected-error>
        </compilation-unit>
      </test-case>
 +    <test-case FilePath="copy-to/negative">
 +      <compilation-unit name="parquet-error-checks">
 +        <output-dir compare="Text">parquet-error-checks</output-dir>
 +        <expected-error>ASX1079: Compilation error: TYPE() Expression is 
required for parquet format</expected-error>
 +        <expected-error>ASX0037: Type mismatch: expected value of type 
integer, but got the value of type BINARY</expected-error>
 +        <expected-error>HYR0132: Extra field in the result, field 'second' 
does not exist at 'nested' in the schema</expected-error>
 +        <expected-error>HYR0131: Result does not follow the schema, group 
type expected but found primitive type at 'nested'</expected-error>
 +        <expected-error>HYR0131: Result does not follow the schema, primitive 
type expected but found group type at 'name'</expected-error>
-         <expected-error>ASX1201: Storage units expected for the field 
'row-group-size' (e.g., 0.1KB, 100kb, 1mb, 3MB, 8.5GB ...). Provided 
'random'</expected-error>
-         <expected-error>ASX1201: Storage units expected for the field 
'page-size' (e.g., 0.1KB, 100kb, 1mb, 3MB, 8.5GB ...). Provided 
'random'</expected-error>
++        <expected-error>ASX1206: Storage units expected for the field 
'row-group-size' (e.g., 0.1KB, 100kb, 1mb, 3MB, 8.5GB ...). Provided 
'random'</expected-error>
++        <expected-error>ASX1206: Storage units expected for the field 
'page-size' (e.g., 0.1KB, 100kb, 1mb, 3MB, 8.5GB ...). Provided 
'random'</expected-error>
 +        <expected-error>ASX1202: Unsupported compression scheme rar. 
Supported schemes for parquet are [gzip, snappy, zstd]</expected-error>
 +        <expected-error>ASX1001: Syntax error</expected-error>
 +        <expected-error>ASX1204: 'binary' type not supported in parquet 
format</expected-error>
 +        <expected-error>ASX1205: Invalid Parquet Writer Version provided '3'. 
Supported values: [1, 2]</expected-error>
 +        <expected-error>ASX1001: Syntax error</expected-error>
 +        <expected-error>ASX1001: Syntax error</expected-error>
 +        <expected-error>ASX1001: Syntax error</expected-error>
 +        </compilation-unit>
 +    </test-case>
      <test-case FilePath="copy-to/negative">
        <compilation-unit name="empty-over">
          <output-dir compare="Text">empty-over</output-dir>
diff --cc 
asterixdb/asterix-common/src/main/java/org/apache/asterix/common/exceptions/ErrorCode.java
index 7c89bc91e6,091fb6774f..ea7cc3bae9
--- 
a/asterixdb/asterix-common/src/main/java/org/apache/asterix/common/exceptions/ErrorCode.java
+++ 
b/asterixdb/asterix-common/src/main/java/org/apache/asterix/common/exceptions/ErrorCode.java
@@@ -303,12 -303,8 +303,12 @@@ public enum ErrorCode implements IErro
      DUPLICATE_FIELD_IN_PRIMARY_KEY(1198),
      INCOMPATIBLE_FIELDS_IN_PRIMARY_KEY(1199),
      PREFIX_SHOULD_NOT_START_WITH_SLASH(1200),
-     ILLEGAL_SIZE_PROVIDED(1201),
+     INVALID_DELTA_TABLE_FORMAT(1201),
 -
 +    UNSUPPORTED_WRITER_COMPRESSION_SCHEME(1202),
 +    INVALID_PARQUET_SCHEMA(1203),
 +    TYPE_UNSUPPORTED_PARQUET_WRITE(1204),
 +    INVALID_PARQUET_WRITER_VERSION(1205),
-     INVALID_DELTA_TABLE_FORMAT(1206),
++    ILLEGAL_SIZE_PROVIDED(1206),
      // Feed errors
      DATAFLOW_ILLEGAL_STATE(3001),
      UTIL_DATAFLOW_UTILS_TUPLE_TOO_LARGE(3002),
diff --cc asterixdb/asterix-common/src/main/resources/asx_errormsg/en.properties
index d1e01f66f9,8e3ccffff3..3e2f124b27
--- a/asterixdb/asterix-common/src/main/resources/asx_errormsg/en.properties
+++ b/asterixdb/asterix-common/src/main/resources/asx_errormsg/en.properties
@@@ -305,12 -305,7 +305,12 @@@
  1198 = Duplicate field '%1$s' in primary key
  1199 = Fields '%1$s' and '%2$s' are incompatible for primary key
  1200 = Prefix should not start with "/". Prefix: '%1$s'
- 1201 = Storage units expected for the field '%1$s' (e.g., 0.1KB, 100kb, 1mb, 
3MB, 8.5GB ...). Provided '%2$s'
+ 1201 = Supported file format for 'delta' tables is 'parquet', but '%1$s' was 
provided.
 +1202 = Unsupported compression scheme %1$s. Supported schemes for %2$s are 
%3$s
 +1203 = Invalid schema provided: '%1$s'
 +1204 = '%1$s' type not supported in parquet format
 +1205 = Invalid Parquet Writer Version provided '%1$s'. Supported values: %2$s
- 1206 = Supported file format for 'delta' tables is 'parquet', but '%1$s' was 
provided.
++1206 = Storage units expected for the field '%1$s' (e.g., 0.1KB, 100kb, 1mb, 
3MB, 8.5GB ...). Provided '%2$s'
  
  # Feed Errors
  3001 = Illegal state.

Reply via email to