Repository: incubator-impala Updated Branches: refs/heads/hadoop-next 0449b5bea -> 75a857c0c
IMPALA-4047: Remove occurrences of 'CDH'/'cdh' from repo This change removes some of the occurrences of the strings 'CDH'/'cdh' from the Impala repository. References to Cloudera-internal Jiras have been replaced with upstream Jira issues on issues.cloudera.org. For several categories of occurrences (e.g. pom.xml files, DOWNLOAD_CDH_COMPONENTS) I also created a list of follow-up Jiras to remove the occurrences left after this change. Change-Id: Icb37e2ef0cd9fa0e581d359c5dd3db7812b7b2c8 Reviewed-on: http://gerrit.cloudera.org:8080/4187 Reviewed-by: Jim Apple <jbap...@cloudera.com> Reviewed-by: Alex Behm <alex.b...@cloudera.com> Tested-by: Internal Jenkins Project: http://git-wip-us.apache.org/repos/asf/incubator-impala/repo Commit: http://git-wip-us.apache.org/repos/asf/incubator-impala/commit/ef4c9958 Tree: http://git-wip-us.apache.org/repos/asf/incubator-impala/tree/ef4c9958 Diff: http://git-wip-us.apache.org/repos/asf/incubator-impala/diff/ef4c9958 Branch: refs/heads/hadoop-next Commit: ef4c9958d018ae0b86cd4318e3288af6a2181c7c Parents: 0449b5b Author: Lars Volker <l...@cloudera.com> Authored: Wed Aug 31 15:12:12 2016 +0200 Committer: Internal Jenkins <cloudera-hud...@gerrit.cloudera.org> Committed: Thu Oct 13 00:40:41 2016 +0000 ---------------------------------------------------------------------- .gitignore | 1 - be/src/exec/hbase-table-scanner.cc | 2 +- be/src/exprs/expr-test.cc | 2 +- be/src/scheduling/simple-scheduler.h | 3 ++- be/src/thirdparty/squeasel/squeasel.h | 4 ++-- be/src/util/redactor.detail.h | 3 +-- be/src/util/redactor.h | 3 +-- bin/build_thirdparty.sh | 1 - bin/load-data.py | 2 +- common/thrift/generate_metrics.py | 2 +- fe/pom.xml | 2 +- fe/src/main/java/org/apache/impala/analysis/SetStmt.java | 2 +- .../main/java/org/apache/impala/catalog/HBaseTable.java | 5 ++--- .../org/apache/impala/catalog/HdfsStorageDescriptor.java | 2 +- .../org/apache/impala/catalog/PartitionStatsUtil.java | 8 ++++---- .../java/org/apache/impala/testutil/ImpalaJdbcClient.java | 3 +-- fe/src/test/resources/postgresql-hive-site.xml.template | 4 ++-- infra/deploy/deploy.py | 4 ++-- testdata/cluster/.gitignore | 1 - testdata/cluster/admin | 7 +++---- testdata/datasets/functional/schema_constraints.csv | 2 +- .../functional-planner/queries/PlannerTest/hdfs.test | 2 +- .../functional-planner/queries/PlannerTest/lineage.test | 10 +++++----- .../queries/QueryTest/hbase-scan-node.test | 4 ++-- tests/comparison/cluster.py | 2 +- tests/comparison/discrepancy_searcher.py | 8 ++++---- tests/comparison/leopard/impala_docker_env.py | 5 +++-- tests/comparison/query.py | 4 ++-- tests/custom_cluster/test_insert_behaviour.py | 2 +- tests/metadata/test_compute_stats.py | 3 ++- tests/metadata/test_hdfs_encryption.py | 2 +- tests/query_test/test_decimal_queries.py | 5 +++-- tests/shell/test_shell_commandline.py | 2 +- tests/stress/concurrent_select.py | 4 ++-- 34 files changed, 56 insertions(+), 60 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/.gitignore ---------------------------------------------------------------------- diff --git a/.gitignore b/.gitignore index 4a7f11a..849ee61 100644 --- a/.gitignore +++ b/.gitignore @@ -40,7 +40,6 @@ tests/build tests/results tests/test-hive-udfs/target/ -cdh-*-hdfs-data/ avro_schemas/ logs/ http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/be/src/exec/hbase-table-scanner.cc ---------------------------------------------------------------------- diff --git a/be/src/exec/hbase-table-scanner.cc b/be/src/exec/hbase-table-scanner.cc index fe8e358..79e45d0 100644 --- a/be/src/exec/hbase-table-scanner.cc +++ b/be/src/exec/hbase-table-scanner.cc @@ -167,7 +167,7 @@ Status HBaseTableScanner::Init() { RETURN_IF_ERROR( JniUtil::GetGlobalClassRef(env, "org/apache/hadoop/hbase/Cell", &cell_cl_)); } else { - // Assume a non-CDH5 HBase version because the Cell class wasn't found. + // Assume a HBase version prior to 0.95.2 because the Cell class wasn't found. LOG(INFO) << "Detected HBase version < 0.95.2"; RETURN_IF_ERROR( JniUtil::GetGlobalClassRef(env, "org/apache/hadoop/hbase/KeyValue", http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/be/src/exprs/expr-test.cc ---------------------------------------------------------------------- diff --git a/be/src/exprs/expr-test.cc b/be/src/exprs/expr-test.cc index 8064de0..88565e1 100644 --- a/be/src/exprs/expr-test.cc +++ b/be/src/exprs/expr-test.cc @@ -3525,7 +3525,7 @@ TEST_F(ExprTest, UnaryOperators) { // TODO: I think a lot of these casts are not necessary and we should fix this TEST_F(ExprTest, TimestampFunctions) { - // Regression test for CDH-19918 + // Regression test for IMPALA-4209 TestStringValue("cast(from_utc_timestamp(cast(1301180400 as timestamp)," "'Europe/Moscow') as string)", "2011-03-27 03:00:00"); TestStringValue("cast(from_utc_timestamp(cast(1301180399 as timestamp)," http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/be/src/scheduling/simple-scheduler.h ---------------------------------------------------------------------- diff --git a/be/src/scheduling/simple-scheduler.h b/be/src/scheduling/simple-scheduler.h index b7cc83e..6a9d6db 100644 --- a/be/src/scheduling/simple-scheduler.h +++ b/be/src/scheduling/simple-scheduler.h @@ -56,7 +56,8 @@ class SchedulerWrapper; /// TODO: Notice when there are duplicate statestore registrations (IMPALA-23) /// TODO: Track assignments (assignment_ctx in ComputeScanRangeAssignment) per query /// instead of per plan node? -/// TODO: Remove disable_cached_reads query option in CDH6 +/// TODO: Remove disable_cached_reads query option in the next compatibility-breaking +/// release (IMPALA-2963) /// TODO: Replace the usage of shared_ptr with atomic_shared_ptr once compilers support /// it. Alternatively consider using Kudu's rw locks. /// TODO: Inject global dependencies into the class (for example ExecEnv::GetInstance(), http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/be/src/thirdparty/squeasel/squeasel.h ---------------------------------------------------------------------- diff --git a/be/src/thirdparty/squeasel/squeasel.h b/be/src/thirdparty/squeasel/squeasel.h index a17f9ef..68a4034 100644 --- a/be/src/thirdparty/squeasel/squeasel.h +++ b/be/src/thirdparty/squeasel/squeasel.h @@ -91,7 +91,7 @@ struct sq_request_info { // This structure needs to be passed to sq_start(), to let squeasel know // which callbacks to invoke. For detailed description, see -// https://github.mtv.cloudera.com/CDH/squeasel/blob/master/UserManual.md +// https://github.com/cloudera/squeasel/blob/master/UserManual.md struct sq_callbacks { // Called when squeasel has received new HTTP request. // If callback returns non-zero, @@ -182,7 +182,7 @@ struct sq_callbacks { // }; // struct sq_context *ctx = sq_start(&my_func, NULL, options); // -// Refer to https://github.mtv.cloudera.com/CDH/squeasel/blob/master/UserManual.md +// Refer to https://github.com/cloudera/squeasel/blob/master/UserManual.md // for the list of valid option and their possible values. // // Return: http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/be/src/util/redactor.detail.h ---------------------------------------------------------------------- diff --git a/be/src/util/redactor.detail.h b/be/src/util/redactor.detail.h index 30b3f71..f0030ad 100644 --- a/be/src/util/redactor.detail.h +++ b/be/src/util/redactor.detail.h @@ -19,8 +19,7 @@ /// such as log files or a web ui. For example, this library could be used to log credit /// card numbers as XXXX-...-XXXX instead of actual numbers. // -/// The java original is http://github.mtv.cloudera.com/CDH/logredactor. -/// TODO: Replace the above link with the public version once it is available. +/// The java original is https://github.com/cloudera/logredactor . #ifndef IMPALA_UTIL_REDACTOR_DETAIL_H_ #define IMPALA_UTIL_REDACTOR_DETAIL_H_ http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/be/src/util/redactor.h ---------------------------------------------------------------------- diff --git a/be/src/util/redactor.h b/be/src/util/redactor.h index a3a0c1d..ce8ed79 100644 --- a/be/src/util/redactor.h +++ b/be/src/util/redactor.h @@ -19,8 +19,7 @@ /// such as log files or a web ui. For example, this library could be used to log credit /// card numbers as XXXX-...-XXXX instead of actual numbers. // -/// The java original is http://github.mtv.cloudera.com/CDH/logredactor. -/// TODO: Replace the above link with the public version once it is available. +/// The java original is https://github.com/cloudera/logredactor . #ifndef IMPALA_UTIL_REDACTOR_H #define IMPALA_UTIL_REDACTOR_H http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/bin/build_thirdparty.sh ---------------------------------------------------------------------- diff --git a/bin/build_thirdparty.sh b/bin/build_thirdparty.sh index 2c965a8..4543225 100755 --- a/bin/build_thirdparty.sh +++ b/bin/build_thirdparty.sh @@ -43,7 +43,6 @@ BUILD_SASL=0 BUILD_LDAP=0 BUILD_SNAPPY=0 BUILD_PPROF=0 -BUILD_CDH4EXTRAS=0 BUILD_LZ4=0 for ARG in $* http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/bin/load-data.py ---------------------------------------------------------------------- diff --git a/bin/load-data.py b/bin/load-data.py index e1bfbbb..df41786 100755 --- a/bin/load-data.py +++ b/bin/load-data.py @@ -85,7 +85,7 @@ AVRO_SCHEMA_DIR = "avro_schemas" GENERATE_SCHEMA_CMD = "generate-schema-statements.py --exploration_strategy=%s "\ "--workload=%s --scale_factor=%s --verbose" -# Load data using Hive's beeline because the Hive shell has regressed (CDH-17222). +# Load data using Hive's beeline because the Hive shell has regressed (HIVE-5515). # The Hive shell is stateful, meaning that certain series of actions lead to problems. # Examples of problems due to the statefullness of the Hive shell: # - Creating an HBase table changes the replication factor to 1 for subsequent LOADs. http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/common/thrift/generate_metrics.py ---------------------------------------------------------------------- diff --git a/common/thrift/generate_metrics.py b/common/thrift/generate_metrics.py index dae92bf..e878a29 100755 --- a/common/thrift/generate_metrics.py +++ b/common/thrift/generate_metrics.py @@ -47,7 +47,7 @@ parser.add_option("--output_mdl_path", dest="output_mdl_path", help="The path of the output mdl file. Default: %default") # TODO: get default version value from bin/save-version.sh parser.add_option("--output_mdl_version", dest="output_mdl_version", - metavar="IMPALA_VERSION", default="2.5.0-cdh5", + metavar="IMPALA_VERSION", default="2.8.0-SNAPSHOT", help="The Impala version that is written in the output mdl.") options, args = parser.parse_args() http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/fe/pom.xml ---------------------------------------------------------------------- diff --git a/fe/pom.xml b/fe/pom.xml index 08f6623..c4e7c44 100644 --- a/fe/pom.xml +++ b/fe/pom.xml @@ -207,7 +207,7 @@ under the License. <version>${env.IMPALA_THRIFT_JAVA_VERSION}</version> <!-- libthrift depends httpcore 4.1.3 which does not work with KMS. To workaround this problem the dependency is excluded here and we explicitly add a newer - httpcore dependency version. See CDH-22024. TODO: Find a better fix. --> + httpcore dependency version. See IMPALA-4210. TODO: Find a better fix. --> <exclusions> <exclusion> <groupId>org.apache.httpcomponents</groupId> http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/fe/src/main/java/org/apache/impala/analysis/SetStmt.java ---------------------------------------------------------------------- diff --git a/fe/src/main/java/org/apache/impala/analysis/SetStmt.java b/fe/src/main/java/org/apache/impala/analysis/SetStmt.java index 87c6727..d307ca7 100644 --- a/fe/src/main/java/org/apache/impala/analysis/SetStmt.java +++ b/fe/src/main/java/org/apache/impala/analysis/SetStmt.java @@ -27,7 +27,7 @@ public class SetStmt extends StatementBase { private final String key_; private final String value_; - // This key is deprecated in CDH5.2; COMPRESSION_CODEC_KEY replaces this + // This key is deprecated in Impala 2.0; COMPRESSION_CODEC_KEY replaces this private static final String DEPRECATED_PARQUET_CODEC_KEY = "PARQUET_COMPRESSION_CODEC"; private static final String COMPRESSION_CODEC_KEY = "COMPRESSION_CODEC"; http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/fe/src/main/java/org/apache/impala/catalog/HBaseTable.java ---------------------------------------------------------------------- diff --git a/fe/src/main/java/org/apache/impala/catalog/HBaseTable.java b/fe/src/main/java/org/apache/impala/catalog/HBaseTable.java index 3930610..2d1fc78 100644 --- a/fe/src/main/java/org/apache/impala/catalog/HBaseTable.java +++ b/fe/src/main/java/org/apache/impala/catalog/HBaseTable.java @@ -126,7 +126,7 @@ public class HBaseTable extends Table { private final static Configuration hbaseConf_ = HBaseConfiguration.create(); // Cached column families. Used primarily for speeding up row stats estimation - // (see CDH-19292). + // (see IMPALA-4211). private HColumnDescriptor[] columnFamilies_ = null; protected HBaseTable(TableId id, org.apache.hadoop.hive.metastore.api.Table msTbl, @@ -715,8 +715,7 @@ public class HBaseTable extends Table { /** * This is copied from org.apache.hadoop.hbase.client.HTable. The only difference is * that it does not use cache when calling getRegionLocation. - * TODO: Remove this function and use HTable.getRegionsInRange when the non-cache - * version has been ported to CDH (DISTRO-477). + * TODO: Remove this function and use HTable.getRegionsInRange (IMPALA-4082). * Get the corresponding regions for an arbitrary range of keys. * <p> * http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/fe/src/main/java/org/apache/impala/catalog/HdfsStorageDescriptor.java ---------------------------------------------------------------------- diff --git a/fe/src/main/java/org/apache/impala/catalog/HdfsStorageDescriptor.java b/fe/src/main/java/org/apache/impala/catalog/HdfsStorageDescriptor.java index 93121c0..ace00c2 100644 --- a/fe/src/main/java/org/apache/impala/catalog/HdfsStorageDescriptor.java +++ b/fe/src/main/java/org/apache/impala/catalog/HdfsStorageDescriptor.java @@ -60,7 +60,7 @@ public class HdfsStorageDescriptor { "org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe", // (rc) "parquet.hive.serde.ParquetHiveSerDe", // (parquet - legacy) // TODO: Verify the following Parquet SerDe works with Impala and add - // support for the new input/output format classes. See CDH-17085. + // support for the new input/output format classes. See IMPALA-4214. "org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe"); // (parquet) private final static Logger LOG = LoggerFactory.getLogger(HdfsStorageDescriptor.class); http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/fe/src/main/java/org/apache/impala/catalog/PartitionStatsUtil.java ---------------------------------------------------------------------- diff --git a/fe/src/main/java/org/apache/impala/catalog/PartitionStatsUtil.java b/fe/src/main/java/org/apache/impala/catalog/PartitionStatsUtil.java index 222f252..8285e9b 100644 --- a/fe/src/main/java/org/apache/impala/catalog/PartitionStatsUtil.java +++ b/fe/src/main/java/org/apache/impala/catalog/PartitionStatsUtil.java @@ -90,10 +90,10 @@ public class PartitionStatsUtil { return; } - // The HMS has a 4k (as of CDH5.2) limit on the length of any parameter string. The - // serialised version of the partition stats is often larger than this. Therefore, we - // naively 'chunk' the byte string into 4k pieces, and store the number of pieces in a - // separate parameter field. + // The HMS has a 4k (as of Hive 0.13, Impala 2.0) limit on the length of any parameter + // string. The serialised version of the partition stats is often larger than this. + // Therefore, we naively 'chunk' the byte string into 4k pieces, and store the number + // of pieces in a separate parameter field. // // The object itself is first serialised by Thrift, and then base-64 encoded to be a // valid string. This inflates its length somewhat; we may want to consider a http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/fe/src/test/java/org/apache/impala/testutil/ImpalaJdbcClient.java ---------------------------------------------------------------------- diff --git a/fe/src/test/java/org/apache/impala/testutil/ImpalaJdbcClient.java b/fe/src/test/java/org/apache/impala/testutil/ImpalaJdbcClient.java index 7176342..1d34ef0 100644 --- a/fe/src/test/java/org/apache/impala/testutil/ImpalaJdbcClient.java +++ b/fe/src/test/java/org/apache/impala/testutil/ImpalaJdbcClient.java @@ -284,8 +284,7 @@ public class ImpalaJdbcClient { float seconds = (endTime - startTime) / 1000F; LOG.info("Returned " + rowCount + " row(s) in " + seconds + "s"); - // TODO: To work around a JDBC driver issue (CDH-10035), make sure the Statement - // is closed after every query. + // Make sure the Statement is closed after every query. client.getStatement().close(); } } http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/fe/src/test/resources/postgresql-hive-site.xml.template ---------------------------------------------------------------------- diff --git a/fe/src/test/resources/postgresql-hive-site.xml.template b/fe/src/test/resources/postgresql-hive-site.xml.template index d25d7fa..c89abca 100644 --- a/fe/src/test/resources/postgresql-hive-site.xml.template +++ b/fe/src/test/resources/postgresql-hive-site.xml.template @@ -167,7 +167,7 @@ </property> --> -<!-- Start CDH-17946 --> +<!-- Start IMPALA-4125 --> <property> <name>hbase.security.authentication</name> <value>kerberos</value> @@ -187,7 +187,7 @@ <name>hbase.zookeeper.quorum</name> <value>localhost</value> </property> -<!-- END CDH-17946 --> +<!-- END IMPALA-4125 --> <property> <name>hbase.coprocessor.region.classes</name> http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/infra/deploy/deploy.py ---------------------------------------------------------------------- diff --git a/infra/deploy/deploy.py b/infra/deploy/deploy.py index c4f40d9..56e3c46 100644 --- a/infra/deploy/deploy.py +++ b/infra/deploy/deploy.py @@ -324,7 +324,7 @@ def create_new_service(api, cluster, new_name, deps, scratch_dirs, master_host): print "Created new role %s" % (new_role_name,) def transform_path(rcg_name, rcg_config_dict, rcg_config_name): - # XXX: Do a better job with paths where the role type is embedded. + # TODO: Do a better job with paths where the role type is embedded. # # e.g. /var/log/impalad/lineage --> /var/log/impalad2/lineage val = rcg_config_dict.get(rcg_config_name, None) @@ -335,7 +335,7 @@ def transform_path(rcg_name, rcg_config_dict, rcg_config_name): return {rcg_config_name : new_val} def transform_port(rcg_name, rcg_config_dict, rcg_config_name): - # XXX: Actually resolve all port conflicts. + # TODO: Actually resolve all port conflicts. val = rcg_config_dict.get(rcg_config_name, None) if not val: raise Exception("Could not get %s config for rcg %s" % http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/testdata/cluster/.gitignore ---------------------------------------------------------------------- diff --git a/testdata/cluster/.gitignore b/testdata/cluster/.gitignore index c362112..a8aeb3e 100644 --- a/testdata/cluster/.gitignore +++ b/testdata/cluster/.gitignore @@ -1,2 +1 @@ -/cdh* /kudu http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/testdata/cluster/admin ---------------------------------------------------------------------- diff --git a/testdata/cluster/admin b/testdata/cluster/admin index 98969ca..bbb4755 100755 --- a/testdata/cluster/admin +++ b/testdata/cluster/admin @@ -20,13 +20,12 @@ # This will create/control/destroy a local hdfs+yarn cluster. # -# The original idea was to run each node on a different loopback address but -# https://jira.cloudera.com/browse/CDH-16602 makes that impossible for now. So all roles -# run on 127.0.0.1, just like the standard mini cluster included with hadoop. The -# difference is with this cluster, each role runs in its own process and has its own +# All roles run on 127.0.0.1, just like the standard mini cluster included with hadoop. +# The difference is with this cluster, each role runs in its own process and has its own # configs. For each node, the layout of the configs, logs, start/stop scripts, etc, is # kept as close as possible to a real cluster. For example, the first node will live in # the dir "cdh-<version>/node-1" and its logs would be at "cdh-<version>/node-1/var/log". +# TODO: Run each node on its own IP address, e.g. 127.0.0.1, 127.0.0.2, and so on. set -euo pipefail trap 'echo Error in $0 at line $LINENO: $(awk "NR == $LINENO" $0)' ERR http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/testdata/datasets/functional/schema_constraints.csv ---------------------------------------------------------------------- diff --git a/testdata/datasets/functional/schema_constraints.csv b/testdata/datasets/functional/schema_constraints.csv index a22a432..4c9da6d 100644 --- a/testdata/datasets/functional/schema_constraints.csv +++ b/testdata/datasets/functional/schema_constraints.csv @@ -123,7 +123,7 @@ table_name:widerow, constraint:exclude, table_format:hbase/none/none table_name:nullformat_custom, constraint:exclude, table_format:hbase/none/none table_name:unsupported_types, constraint:exclude, table_format:hbase/none/none -# On CDH4, decimal can only be tested on formats Impala can write to (text and parquet) +# Decimal can only be tested on formats Impala can write to (text and parquet). # TODO: add Avro once Hive or Impala can write Avro decimals table_name:decimal_tbl, constraint:restrict_to, table_format:text/none/none table_name:decimal_tiny, constraint:restrict_to, table_format:text/none/none http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/testdata/workloads/functional-planner/queries/PlannerTest/hdfs.test ---------------------------------------------------------------------- diff --git a/testdata/workloads/functional-planner/queries/PlannerTest/hdfs.test b/testdata/workloads/functional-planner/queries/PlannerTest/hdfs.test index 880017b..785d111 100644 --- a/testdata/workloads/functional-planner/queries/PlannerTest/hdfs.test +++ b/testdata/workloads/functional-planner/queries/PlannerTest/hdfs.test @@ -623,7 +623,7 @@ NODE 0: HDFS SPLIT hdfs://localhost:20500/test-warehouse/alltypesagg/year=2010/month=1/day=8/100108.txt 0:76263 HDFS SPLIT hdfs://localhost:20500/test-warehouse/alltypesagg/year=2010/month=1/day=__HIVE_DEFAULT_PARTITION__/000000_0 0:72759 ==== -# Predicates on a partition key with no values (see CDH-20089) +# Predicates on a partition key with no values (see IMPALA-4128). select * from functional.emptytable where f2 = 10 ---- PLAN 00:SCAN HDFS [functional.emptytable] http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/testdata/workloads/functional-planner/queries/PlannerTest/lineage.test ---------------------------------------------------------------------- diff --git a/testdata/workloads/functional-planner/queries/PlannerTest/lineage.test b/testdata/workloads/functional-planner/queries/PlannerTest/lineage.test index 9ba2ab7..20c117e 100644 --- a/testdata/workloads/functional-planner/queries/PlannerTest/lineage.test +++ b/testdata/workloads/functional-planner/queries/PlannerTest/lineage.test @@ -3435,7 +3435,7 @@ from ] } ==== -# Create view statement (CDH-26149) +# Create view statement (IMPALA-4219) create view test_view_lineage as select id from functional.alltypestiny ---- LINEAGE { @@ -3468,7 +3468,7 @@ create view test_view_lineage as select id from functional.alltypestiny ] } ==== -# Create view with nested inline views (CDH-26149) +# Create view with nested inline views (IMPALA-4219) create view test_view_lineage (a, b) as select c1, x2 from ( select c1, min(c2) x2 @@ -3547,7 +3547,7 @@ limit 0 ] } ==== -# Create view with complex select statement (CDH-26149) +# Create view with complex select statement (IMPALA-4219) create view test_view_lineage (a1, a2, a3, a4, a5, a6, a7) as select x.smallint_col, x.id, x.tinyint_col, c.id, x.int_col, x.float_col, c.string_col from functional.alltypessmall c @@ -3734,7 +3734,7 @@ create view test_view_lineage (a1, a2, a3, a4, a5, a6, a7) as ] } ==== -# Create view with select statement containing an analytic function (CDH-26149) +# Create view with select statement containing an analytic function (IMPALA-4219) create view test_view_lineage as select * from ( select sum(a.tinyint_col) over (partition by a.smallint_col order by a.id), @@ -3878,7 +3878,7 @@ create view test_view_lineage as ] } ==== -# Alter view statement (CDH-26149) +# Alter view statement (IMPALA-4219) alter view functional.alltypes_view as select id from functional.alltypestiny ---- LINEAGE { http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/testdata/workloads/functional-query/queries/QueryTest/hbase-scan-node.test ---------------------------------------------------------------------- diff --git a/testdata/workloads/functional-query/queries/QueryTest/hbase-scan-node.test b/testdata/workloads/functional-query/queries/QueryTest/hbase-scan-node.test index f23fa56..55415eb 100644 --- a/testdata/workloads/functional-query/queries/QueryTest/hbase-scan-node.test +++ b/testdata/workloads/functional-query/queries/QueryTest/hbase-scan-node.test @@ -636,7 +636,7 @@ select * from alltypessmallbinary INT, BIGINT, BOOLEAN, STRING, DOUBLE, FLOAT, INT, INT, SMALLINT, STRING, TIMESTAMP, TINYINT, INT ==== ---- QUERY -# Scan an HBase table with multiple column families (CDH-18969) +# Scan an HBase table with multiple column families (IMPALA-4220) select * from hbasecolumnfamilies ---- RESULTS 0,true,0,0,0,0,'01/01/09',0,0,'0',2009-01-01 00:00:00 @@ -649,4 +649,4 @@ select * from hbasecolumnfamilies 7,false,1,1,1,10,'04/01/09',10.1,1.100000023841858,'1',2009-04-01 00:01:00 ---- TYPES INT, BOOLEAN, TINYINT, SMALLINT, INT, BIGINT, STRING, DOUBLE, FLOAT, STRING, TIMESTAMP -==== \ No newline at end of file +==== http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/tests/comparison/cluster.py ---------------------------------------------------------------------- diff --git a/tests/comparison/cluster.py b/tests/comparison/cluster.py index bfb8fa4..ab1e4f3 100644 --- a/tests/comparison/cluster.py +++ b/tests/comparison/cluster.py @@ -630,7 +630,7 @@ class Impalad(object): try: self._request_web_page("/cancel_query", params={"query_id": id}) except requests.exceptions.HTTPError as e: - # XXX: Handle losing the race + # TODO: Handle losing the race raise e def shell(self, cmd, timeout_secs=DEFAULT_TIMEOUT): http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/tests/comparison/discrepancy_searcher.py ---------------------------------------------------------------------- diff --git a/tests/comparison/discrepancy_searcher.py b/tests/comparison/discrepancy_searcher.py index a1e60e2..57e284a 100755 --- a/tests/comparison/discrepancy_searcher.py +++ b/tests/comparison/discrepancy_searcher.py @@ -247,7 +247,7 @@ class QueryResultComparator(object): class QueryExecutor(object): '''Concurrently executes queries''' - # XXX: Set to false while IMPALA-3336 is a problem. Disabling random query options + # TODO: Set to false while IMPALA-3336 is a problem. Disabling random query options # seems to reduce IMPALA-3336 occurances. ENABLE_RANDOM_QUERY_OPTIONS = False @@ -379,9 +379,9 @@ class QueryExecutor(object): cursor.conn.kill() LOG.debug('Kill connection') try: - # XXX: Sometimes this takes a very long time causing the program to appear to - # hang. Maybe this should be done in another thread so a timeout can be - # applied? + # TODO: Sometimes this takes a very long time causing the program to appear to + # hang. Maybe this should be done in another thread so a timeout can be + # applied? cursor.close() except Exception as e: LOG.info('Error closing cursor: %s', e) http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/tests/comparison/leopard/impala_docker_env.py ---------------------------------------------------------------------- diff --git a/tests/comparison/leopard/impala_docker_env.py b/tests/comparison/leopard/impala_docker_env.py index 3f1d292..715ad65 100755 --- a/tests/comparison/leopard/impala_docker_env.py +++ b/tests/comparison/leopard/impala_docker_env.py @@ -32,7 +32,7 @@ import os IMPALA_HOME = '/home/dev/Impala' CORE_PATH = '/tmp/core_files' DEFAULT_BRANCH_NAME = 'origin/cdh5-trunk' -DEFAULT_DOCKER_IMAGE_NAME = 'impala-desktop.ca.cloudera.com:5000/ubuntu-14.04:cdh5-trunk' +DEFAULT_DOCKER_IMAGE_NAME = 'cloudera/impala-dev' DOCKER_USER_NAME = 'dev' NUM_START_ATTEMPTS = 50 NUM_FABRIC_ATTEMPTS = 50 @@ -152,10 +152,11 @@ class ImpalaDockerEnv(object): elif SHOULD_BUILD_IMPALA: build_command = ( 'docker-boot && cd {IMPALA_HOME} ' - '&& git fetch --all && git checkout origin/cdh5-trunk ' + '&& git fetch --all && git checkout DEFAULT_BRANCH_NAME ' '&& source {IMPALA_HOME}/bin/impala-config.sh ' '&& {IMPALA_HOME}/buildall.sh -notests').format( IMPALA_HOME = IMPALA_HOME, + DEFAULT_BRANCH_NAME = DEFAULT_BRANCH_NAME, CORE_PATH = CORE_PATH) if build_command: http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/tests/comparison/query.py ---------------------------------------------------------------------- diff --git a/tests/comparison/query.py b/tests/comparison/query.py index b928bf3..f6ec65c 100644 --- a/tests/comparison/query.py +++ b/tests/comparison/query.py @@ -336,8 +336,8 @@ class Subquery(ValExpr): result set. ''' - # XXX: So far it seems fine to use this class for both scalar/non scalar cases but - # this could lead to unexpected behavior or be a silent cause of problems... + # TODO: So far it seems fine to use this class for both scalar/non scalar cases but + # this could lead to unexpected behavior or be a silent cause of problems... def __init__(self, query): self.query = query http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/tests/custom_cluster/test_insert_behaviour.py ---------------------------------------------------------------------- diff --git a/tests/custom_cluster/test_insert_behaviour.py b/tests/custom_cluster/test_insert_behaviour.py index 9cec88f..6df8ed7 100644 --- a/tests/custom_cluster/test_insert_behaviour.py +++ b/tests/custom_cluster/test_insert_behaviour.py @@ -120,7 +120,7 @@ class TestInsertBehaviourCustomCluster(CustomClusterTestSuite): self.execute_query_expect_success(client, "INSERT INTO %s" " PARTITION(p1=1, p2=3, p3=4) VALUES(1)" % TEST_TBL) # Would be 777 if inheritance was enabled - if not IS_ISILON: # CDH-27688 + if not IS_ISILON: # IMPALA-4221 self._check_partition_perms("p1=1/p2=3/", default_perms) self._check_partition_perms("p1=1/p2=3/p3=4/", default_perms) finally: http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/tests/metadata/test_compute_stats.py ---------------------------------------------------------------------- diff --git a/tests/metadata/test_compute_stats.py b/tests/metadata/test_compute_stats.py index 6f6d31a..9ecdcdf 100644 --- a/tests/metadata/test_compute_stats.py +++ b/tests/metadata/test_compute_stats.py @@ -42,7 +42,8 @@ class TestComputeStats(ImpalaTestSuite): @SkipIfLocal.hdfs_blocks def test_compute_stats(self, vector, unique_database): self.run_test_case('QueryTest/compute-stats', vector, unique_database) - # Test compute stats on decimal columns separately so we can vary between CDH4/5 + # Test compute stats on decimal columns separately so we can vary between platforms + # with and without write support for decimals (Hive < 0.11 and >= 0.11). self.run_test_case('QueryTest/compute-stats-decimal', vector, unique_database) def test_compute_stats_incremental(self, vector, unique_database): http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/tests/metadata/test_hdfs_encryption.py ---------------------------------------------------------------------- diff --git a/tests/metadata/test_hdfs_encryption.py b/tests/metadata/test_hdfs_encryption.py index 196b2ea..a257de0 100644 --- a/tests/metadata/test_hdfs_encryption.py +++ b/tests/metadata/test_hdfs_encryption.py @@ -146,7 +146,7 @@ class TestHdfsEncryption(ImpalaTestSuite): def test_drop_partition_encrypt(self): """Verifies if alter <tbl> drop partition purge works in case where the Trash dir and partition dir are in different encryption - zones. Check CDH-31350 for details""" + zones. Check IMPALA-2310 for details""" self.client.execute("create table {0}.t1(i int) partitioned\ by (j int)".format(TEST_DB)) # Add three partitions (j=1), (j=2), (j=3) to table t1 http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/tests/query_test/test_decimal_queries.py ---------------------------------------------------------------------- diff --git a/tests/query_test/test_decimal_queries.py b/tests/query_test/test_decimal_queries.py index 0404ea1..f95d940 100644 --- a/tests/query_test/test_decimal_queries.py +++ b/tests/query_test/test_decimal_queries.py @@ -35,8 +35,9 @@ class TestDecimalQueries(ImpalaTestSuite): cls.TestMatrix.add_dimension( TestDimension('batch_size', *TestDecimalQueries.BATCH_SIZES)) - # On CDH4, hive does not support decimal so we can't run these tests against - # the other file formats. Enable them on C5. + # Hive < 0.11 does not support decimal so we can't run these tests against the other + # file formats. + # TODO: Enable them on Hive >= 0.11. cls.TestMatrix.add_constraint(lambda v:\ (v.get_value('table_format').file_format == 'text' and v.get_value('table_format').compression_codec == 'none') or http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/tests/shell/test_shell_commandline.py ---------------------------------------------------------------------- diff --git a/tests/shell/test_shell_commandline.py b/tests/shell/test_shell_commandline.py index 31cf570..16a68ae 100644 --- a/tests/shell/test_shell_commandline.py +++ b/tests/shell/test_shell_commandline.py @@ -200,7 +200,7 @@ class TestImpalaShell(ImpalaTestSuite): assert 'Problem parsing file' in result.stderr def test_no_warnings_in_log_with_quiet_mode(self): - """Regression test for CDH-21036.""" + """Regression test for IMPALA-4222.""" args = ('-q "set abort_on_error=false;' ' select count(*) from functional_seq_snap.bad_seq_snap" --quiet') result = run_impala_shell_cmd(args) http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/tests/stress/concurrent_select.py ---------------------------------------------------------------------- diff --git a/tests/stress/concurrent_select.py b/tests/stress/concurrent_select.py index 1d1d142..b76d713 100755 --- a/tests/stress/concurrent_select.py +++ b/tests/stress/concurrent_select.py @@ -335,8 +335,8 @@ class StressRunner(object): If a query completes without error, the result will be verified. An error will be raised upon a result mismatch. """ - # XXX: The state from a previous run should be cleared out. This isn't really a - # problem now because the one caller (main()) never calls a second time. + # TODO: The state from a previous run should be cleared out. This isn't really a + # problem now because the one caller (main()) never calls a second time. if self.startup_queries_per_sec <= 0: raise Exception("Startup queries per second must be positive")