Repository: incubator-impala
Updated Branches:
  refs/heads/hadoop-next 0449b5bea -> 75a857c0c


IMPALA-4047: Remove occurrences of 'CDH'/'cdh' from repo

This change removes some of the occurrences of the strings 'CDH'/'cdh'
from the Impala repository. References to Cloudera-internal Jiras have
been replaced with upstream Jira issues on issues.cloudera.org.

For several categories of occurrences (e.g. pom.xml files,
DOWNLOAD_CDH_COMPONENTS) I also created a list of follow-up Jiras to
remove the occurrences left after this change.

Change-Id: Icb37e2ef0cd9fa0e581d359c5dd3db7812b7b2c8
Reviewed-on: http://gerrit.cloudera.org:8080/4187
Reviewed-by: Jim Apple <jbap...@cloudera.com>
Reviewed-by: Alex Behm <alex.b...@cloudera.com>
Tested-by: Internal Jenkins


Project: http://git-wip-us.apache.org/repos/asf/incubator-impala/repo
Commit: http://git-wip-us.apache.org/repos/asf/incubator-impala/commit/ef4c9958
Tree: http://git-wip-us.apache.org/repos/asf/incubator-impala/tree/ef4c9958
Diff: http://git-wip-us.apache.org/repos/asf/incubator-impala/diff/ef4c9958

Branch: refs/heads/hadoop-next
Commit: ef4c9958d018ae0b86cd4318e3288af6a2181c7c
Parents: 0449b5b
Author: Lars Volker <l...@cloudera.com>
Authored: Wed Aug 31 15:12:12 2016 +0200
Committer: Internal Jenkins <cloudera-hud...@gerrit.cloudera.org>
Committed: Thu Oct 13 00:40:41 2016 +0000

----------------------------------------------------------------------
 .gitignore                                                |  1 -
 be/src/exec/hbase-table-scanner.cc                        |  2 +-
 be/src/exprs/expr-test.cc                                 |  2 +-
 be/src/scheduling/simple-scheduler.h                      |  3 ++-
 be/src/thirdparty/squeasel/squeasel.h                     |  4 ++--
 be/src/util/redactor.detail.h                             |  3 +--
 be/src/util/redactor.h                                    |  3 +--
 bin/build_thirdparty.sh                                   |  1 -
 bin/load-data.py                                          |  2 +-
 common/thrift/generate_metrics.py                         |  2 +-
 fe/pom.xml                                                |  2 +-
 fe/src/main/java/org/apache/impala/analysis/SetStmt.java  |  2 +-
 .../main/java/org/apache/impala/catalog/HBaseTable.java   |  5 ++---
 .../org/apache/impala/catalog/HdfsStorageDescriptor.java  |  2 +-
 .../org/apache/impala/catalog/PartitionStatsUtil.java     |  8 ++++----
 .../java/org/apache/impala/testutil/ImpalaJdbcClient.java |  3 +--
 fe/src/test/resources/postgresql-hive-site.xml.template   |  4 ++--
 infra/deploy/deploy.py                                    |  4 ++--
 testdata/cluster/.gitignore                               |  1 -
 testdata/cluster/admin                                    |  7 +++----
 testdata/datasets/functional/schema_constraints.csv       |  2 +-
 .../functional-planner/queries/PlannerTest/hdfs.test      |  2 +-
 .../functional-planner/queries/PlannerTest/lineage.test   | 10 +++++-----
 .../queries/QueryTest/hbase-scan-node.test                |  4 ++--
 tests/comparison/cluster.py                               |  2 +-
 tests/comparison/discrepancy_searcher.py                  |  8 ++++----
 tests/comparison/leopard/impala_docker_env.py             |  5 +++--
 tests/comparison/query.py                                 |  4 ++--
 tests/custom_cluster/test_insert_behaviour.py             |  2 +-
 tests/metadata/test_compute_stats.py                      |  3 ++-
 tests/metadata/test_hdfs_encryption.py                    |  2 +-
 tests/query_test/test_decimal_queries.py                  |  5 +++--
 tests/shell/test_shell_commandline.py                     |  2 +-
 tests/stress/concurrent_select.py                         |  4 ++--
 34 files changed, 56 insertions(+), 60 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/.gitignore
----------------------------------------------------------------------
diff --git a/.gitignore b/.gitignore
index 4a7f11a..849ee61 100644
--- a/.gitignore
+++ b/.gitignore
@@ -40,7 +40,6 @@ tests/build
 tests/results
 tests/test-hive-udfs/target/
 
-cdh-*-hdfs-data/
 avro_schemas/
 logs/
 

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/be/src/exec/hbase-table-scanner.cc
----------------------------------------------------------------------
diff --git a/be/src/exec/hbase-table-scanner.cc 
b/be/src/exec/hbase-table-scanner.cc
index fe8e358..79e45d0 100644
--- a/be/src/exec/hbase-table-scanner.cc
+++ b/be/src/exec/hbase-table-scanner.cc
@@ -167,7 +167,7 @@ Status HBaseTableScanner::Init() {
     RETURN_IF_ERROR(
         JniUtil::GetGlobalClassRef(env, "org/apache/hadoop/hbase/Cell", 
&cell_cl_));
   } else {
-    // Assume a non-CDH5 HBase version because the Cell class wasn't found.
+    // Assume a HBase version prior to 0.95.2 because the Cell class wasn't 
found.
     LOG(INFO) << "Detected HBase version < 0.95.2";
     RETURN_IF_ERROR(
         JniUtil::GetGlobalClassRef(env, "org/apache/hadoop/hbase/KeyValue",

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/be/src/exprs/expr-test.cc
----------------------------------------------------------------------
diff --git a/be/src/exprs/expr-test.cc b/be/src/exprs/expr-test.cc
index 8064de0..88565e1 100644
--- a/be/src/exprs/expr-test.cc
+++ b/be/src/exprs/expr-test.cc
@@ -3525,7 +3525,7 @@ TEST_F(ExprTest, UnaryOperators) {
 
 // TODO: I think a lot of these casts are not necessary and we should fix this
 TEST_F(ExprTest, TimestampFunctions) {
-  // Regression test for CDH-19918
+  // Regression test for IMPALA-4209
   TestStringValue("cast(from_utc_timestamp(cast(1301180400 as timestamp),"
       "'Europe/Moscow') as string)", "2011-03-27 03:00:00");
   TestStringValue("cast(from_utc_timestamp(cast(1301180399 as timestamp),"

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/be/src/scheduling/simple-scheduler.h
----------------------------------------------------------------------
diff --git a/be/src/scheduling/simple-scheduler.h 
b/be/src/scheduling/simple-scheduler.h
index b7cc83e..6a9d6db 100644
--- a/be/src/scheduling/simple-scheduler.h
+++ b/be/src/scheduling/simple-scheduler.h
@@ -56,7 +56,8 @@ class SchedulerWrapper;
 /// TODO: Notice when there are duplicate statestore registrations (IMPALA-23)
 /// TODO: Track assignments (assignment_ctx in ComputeScanRangeAssignment) per 
query
 ///       instead of per plan node?
-/// TODO: Remove disable_cached_reads query option in CDH6
+/// TODO: Remove disable_cached_reads query option in the next 
compatibility-breaking
+///       release (IMPALA-2963)
 /// TODO: Replace the usage of shared_ptr with atomic_shared_ptr once 
compilers support
 ///       it. Alternatively consider using Kudu's rw locks.
 /// TODO: Inject global dependencies into the class (for example 
ExecEnv::GetInstance(),

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/be/src/thirdparty/squeasel/squeasel.h
----------------------------------------------------------------------
diff --git a/be/src/thirdparty/squeasel/squeasel.h 
b/be/src/thirdparty/squeasel/squeasel.h
index a17f9ef..68a4034 100644
--- a/be/src/thirdparty/squeasel/squeasel.h
+++ b/be/src/thirdparty/squeasel/squeasel.h
@@ -91,7 +91,7 @@ struct sq_request_info {
 
 // This structure needs to be passed to sq_start(), to let squeasel know
 // which callbacks to invoke. For detailed description, see
-// https://github.mtv.cloudera.com/CDH/squeasel/blob/master/UserManual.md
+// https://github.com/cloudera/squeasel/blob/master/UserManual.md
 struct sq_callbacks {
   // Called when squeasel has received new HTTP request.
   // If callback returns non-zero,
@@ -182,7 +182,7 @@ struct sq_callbacks {
 //   };
 //   struct sq_context *ctx = sq_start(&my_func, NULL, options);
 //
-// Refer to 
https://github.mtv.cloudera.com/CDH/squeasel/blob/master/UserManual.md
+// Refer to https://github.com/cloudera/squeasel/blob/master/UserManual.md
 // for the list of valid option and their possible values.
 //
 // Return:

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/be/src/util/redactor.detail.h
----------------------------------------------------------------------
diff --git a/be/src/util/redactor.detail.h b/be/src/util/redactor.detail.h
index 30b3f71..f0030ad 100644
--- a/be/src/util/redactor.detail.h
+++ b/be/src/util/redactor.detail.h
@@ -19,8 +19,7 @@
 /// such as log files or a web ui. For example, this library could be used to 
log credit
 /// card numbers as XXXX-...-XXXX instead of actual numbers.
 //
-/// The java original is http://github.mtv.cloudera.com/CDH/logredactor.
-/// TODO: Replace the above link with the public version once it is available.
+/// The java original is https://github.com/cloudera/logredactor .
 
 #ifndef IMPALA_UTIL_REDACTOR_DETAIL_H_
 #define IMPALA_UTIL_REDACTOR_DETAIL_H_

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/be/src/util/redactor.h
----------------------------------------------------------------------
diff --git a/be/src/util/redactor.h b/be/src/util/redactor.h
index a3a0c1d..ce8ed79 100644
--- a/be/src/util/redactor.h
+++ b/be/src/util/redactor.h
@@ -19,8 +19,7 @@
 /// such as log files or a web ui. For example, this library could be used to 
log credit
 /// card numbers as XXXX-...-XXXX instead of actual numbers.
 //
-/// The java original is http://github.mtv.cloudera.com/CDH/logredactor.
-/// TODO: Replace the above link with the public version once it is available.
+/// The java original is https://github.com/cloudera/logredactor .
 
 #ifndef IMPALA_UTIL_REDACTOR_H
 #define IMPALA_UTIL_REDACTOR_H

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/bin/build_thirdparty.sh
----------------------------------------------------------------------
diff --git a/bin/build_thirdparty.sh b/bin/build_thirdparty.sh
index 2c965a8..4543225 100755
--- a/bin/build_thirdparty.sh
+++ b/bin/build_thirdparty.sh
@@ -43,7 +43,6 @@ BUILD_SASL=0
 BUILD_LDAP=0
 BUILD_SNAPPY=0
 BUILD_PPROF=0
-BUILD_CDH4EXTRAS=0
 BUILD_LZ4=0
 
 for ARG in $*

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/bin/load-data.py
----------------------------------------------------------------------
diff --git a/bin/load-data.py b/bin/load-data.py
index e1bfbbb..df41786 100755
--- a/bin/load-data.py
+++ b/bin/load-data.py
@@ -85,7 +85,7 @@ AVRO_SCHEMA_DIR = "avro_schemas"
 
 GENERATE_SCHEMA_CMD = "generate-schema-statements.py --exploration_strategy=%s 
"\
                       "--workload=%s --scale_factor=%s --verbose"
-# Load data using Hive's beeline because the Hive shell has regressed 
(CDH-17222).
+# Load data using Hive's beeline because the Hive shell has regressed 
(HIVE-5515).
 # The Hive shell is stateful, meaning that certain series of actions lead to 
problems.
 # Examples of problems due to the statefullness of the Hive shell:
 # - Creating an HBase table changes the replication factor to 1 for subsequent 
LOADs.

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/common/thrift/generate_metrics.py
----------------------------------------------------------------------
diff --git a/common/thrift/generate_metrics.py 
b/common/thrift/generate_metrics.py
index dae92bf..e878a29 100755
--- a/common/thrift/generate_metrics.py
+++ b/common/thrift/generate_metrics.py
@@ -47,7 +47,7 @@ parser.add_option("--output_mdl_path", dest="output_mdl_path",
                   help="The path of the output mdl file. Default: %default")
 # TODO: get default version value from bin/save-version.sh
 parser.add_option("--output_mdl_version", dest="output_mdl_version",
-                  metavar="IMPALA_VERSION", default="2.5.0-cdh5",
+                  metavar="IMPALA_VERSION", default="2.8.0-SNAPSHOT",
                   help="The Impala version that is written in the output mdl.")
 
 options, args = parser.parse_args()

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/fe/pom.xml
----------------------------------------------------------------------
diff --git a/fe/pom.xml b/fe/pom.xml
index 08f6623..c4e7c44 100644
--- a/fe/pom.xml
+++ b/fe/pom.xml
@@ -207,7 +207,7 @@ under the License.
       <version>${env.IMPALA_THRIFT_JAVA_VERSION}</version>
       <!-- libthrift depends httpcore 4.1.3 which does not work with KMS. To 
workaround
            this problem the dependency is excluded here and we explicitly add 
a newer
-           httpcore dependency version. See CDH-22024. TODO: Find a better 
fix. -->
+           httpcore dependency version. See IMPALA-4210. TODO: Find a better 
fix. -->
       <exclusions>
         <exclusion>
           <groupId>org.apache.httpcomponents</groupId>

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/fe/src/main/java/org/apache/impala/analysis/SetStmt.java
----------------------------------------------------------------------
diff --git a/fe/src/main/java/org/apache/impala/analysis/SetStmt.java 
b/fe/src/main/java/org/apache/impala/analysis/SetStmt.java
index 87c6727..d307ca7 100644
--- a/fe/src/main/java/org/apache/impala/analysis/SetStmt.java
+++ b/fe/src/main/java/org/apache/impala/analysis/SetStmt.java
@@ -27,7 +27,7 @@ public class SetStmt extends StatementBase {
   private final String key_;
   private final String value_;
 
-  // This key is deprecated in CDH5.2; COMPRESSION_CODEC_KEY replaces this
+  // This key is deprecated in Impala 2.0; COMPRESSION_CODEC_KEY replaces this
   private static final String DEPRECATED_PARQUET_CODEC_KEY = 
"PARQUET_COMPRESSION_CODEC";
   private static final String COMPRESSION_CODEC_KEY = "COMPRESSION_CODEC";
 

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/fe/src/main/java/org/apache/impala/catalog/HBaseTable.java
----------------------------------------------------------------------
diff --git a/fe/src/main/java/org/apache/impala/catalog/HBaseTable.java 
b/fe/src/main/java/org/apache/impala/catalog/HBaseTable.java
index 3930610..2d1fc78 100644
--- a/fe/src/main/java/org/apache/impala/catalog/HBaseTable.java
+++ b/fe/src/main/java/org/apache/impala/catalog/HBaseTable.java
@@ -126,7 +126,7 @@ public class HBaseTable extends Table {
   private final static Configuration hbaseConf_ = HBaseConfiguration.create();
 
   // Cached column families. Used primarily for speeding up row stats 
estimation
-  // (see CDH-19292).
+  // (see IMPALA-4211).
   private HColumnDescriptor[] columnFamilies_ = null;
 
   protected HBaseTable(TableId id, org.apache.hadoop.hive.metastore.api.Table 
msTbl,
@@ -715,8 +715,7 @@ public class HBaseTable extends Table {
   /**
    * This is copied from org.apache.hadoop.hbase.client.HTable. The only 
difference is
    * that it does not use cache when calling getRegionLocation.
-   * TODO: Remove this function and use HTable.getRegionsInRange when the 
non-cache
-   * version has been ported to CDH (DISTRO-477).
+   * TODO: Remove this function and use HTable.getRegionsInRange (IMPALA-4082).
    * Get the corresponding regions for an arbitrary range of keys.
    * <p>
    *

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/fe/src/main/java/org/apache/impala/catalog/HdfsStorageDescriptor.java
----------------------------------------------------------------------
diff --git 
a/fe/src/main/java/org/apache/impala/catalog/HdfsStorageDescriptor.java 
b/fe/src/main/java/org/apache/impala/catalog/HdfsStorageDescriptor.java
index 93121c0..ace00c2 100644
--- a/fe/src/main/java/org/apache/impala/catalog/HdfsStorageDescriptor.java
+++ b/fe/src/main/java/org/apache/impala/catalog/HdfsStorageDescriptor.java
@@ -60,7 +60,7 @@ public class HdfsStorageDescriptor {
       "org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe", // (rc)
       "parquet.hive.serde.ParquetHiveSerDe", // (parquet - legacy)
       // TODO: Verify the following Parquet SerDe works with Impala and add
-      // support for the new input/output format classes. See CDH-17085.
+      // support for the new input/output format classes. See IMPALA-4214.
       "org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe"); // 
(parquet)
 
   private final static Logger LOG = 
LoggerFactory.getLogger(HdfsStorageDescriptor.class);

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/fe/src/main/java/org/apache/impala/catalog/PartitionStatsUtil.java
----------------------------------------------------------------------
diff --git a/fe/src/main/java/org/apache/impala/catalog/PartitionStatsUtil.java 
b/fe/src/main/java/org/apache/impala/catalog/PartitionStatsUtil.java
index 222f252..8285e9b 100644
--- a/fe/src/main/java/org/apache/impala/catalog/PartitionStatsUtil.java
+++ b/fe/src/main/java/org/apache/impala/catalog/PartitionStatsUtil.java
@@ -90,10 +90,10 @@ public class PartitionStatsUtil {
       return;
     }
 
-    // The HMS has a 4k (as of CDH5.2) limit on the length of any parameter 
string. The
-    // serialised version of the partition stats is often larger than this. 
Therefore, we
-    // naively 'chunk' the byte string into 4k pieces, and store the number of 
pieces in a
-    // separate parameter field.
+    // The HMS has a 4k (as of Hive 0.13, Impala 2.0) limit on the length of 
any parameter
+    // string.  The serialised version of the partition stats is often larger 
than this.
+    // Therefore, we naively 'chunk' the byte string into 4k pieces, and store 
the number
+    // of pieces in a separate parameter field.
     //
     // The object itself is first serialised by Thrift, and then base-64 
encoded to be a
     // valid string. This inflates its length somewhat; we may want to 
consider a

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/fe/src/test/java/org/apache/impala/testutil/ImpalaJdbcClient.java
----------------------------------------------------------------------
diff --git a/fe/src/test/java/org/apache/impala/testutil/ImpalaJdbcClient.java 
b/fe/src/test/java/org/apache/impala/testutil/ImpalaJdbcClient.java
index 7176342..1d34ef0 100644
--- a/fe/src/test/java/org/apache/impala/testutil/ImpalaJdbcClient.java
+++ b/fe/src/test/java/org/apache/impala/testutil/ImpalaJdbcClient.java
@@ -284,8 +284,7 @@ public class ImpalaJdbcClient {
       float seconds = (endTime - startTime) / 1000F;
       LOG.info("Returned " + rowCount + " row(s) in " + seconds + "s");
 
-      // TODO: To work around a JDBC driver issue (CDH-10035), make sure the 
Statement
-      // is closed after every query.
+      // Make sure the Statement is closed after every query.
       client.getStatement().close();
     }
   }

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/fe/src/test/resources/postgresql-hive-site.xml.template
----------------------------------------------------------------------
diff --git a/fe/src/test/resources/postgresql-hive-site.xml.template 
b/fe/src/test/resources/postgresql-hive-site.xml.template
index d25d7fa..c89abca 100644
--- a/fe/src/test/resources/postgresql-hive-site.xml.template
+++ b/fe/src/test/resources/postgresql-hive-site.xml.template
@@ -167,7 +167,7 @@
 </property>
 -->
 
-<!-- Start CDH-17946 -->
+<!-- Start IMPALA-4125 -->
 <property>
   <name>hbase.security.authentication</name>
   <value>kerberos</value>
@@ -187,7 +187,7 @@
   <name>hbase.zookeeper.quorum</name>
   <value>localhost</value>
 </property>
-<!-- END CDH-17946 -->
+<!-- END IMPALA-4125 -->
 
 <property>
   <name>hbase.coprocessor.region.classes</name>

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/infra/deploy/deploy.py
----------------------------------------------------------------------
diff --git a/infra/deploy/deploy.py b/infra/deploy/deploy.py
index c4f40d9..56e3c46 100644
--- a/infra/deploy/deploy.py
+++ b/infra/deploy/deploy.py
@@ -324,7 +324,7 @@ def create_new_service(api, cluster, new_name, deps, 
scratch_dirs, master_host):
             print "Created new role %s" % (new_role_name,)
 
 def transform_path(rcg_name, rcg_config_dict, rcg_config_name):
-    # XXX: Do a better job with paths where the role type is embedded.
+    # TODO: Do a better job with paths where the role type is embedded.
     #
     # e.g. /var/log/impalad/lineage --> /var/log/impalad2/lineage
     val = rcg_config_dict.get(rcg_config_name, None)
@@ -335,7 +335,7 @@ def transform_path(rcg_name, rcg_config_dict, 
rcg_config_name):
     return {rcg_config_name : new_val}
 
 def transform_port(rcg_name, rcg_config_dict, rcg_config_name):
-    # XXX: Actually resolve all port conflicts.
+    # TODO: Actually resolve all port conflicts.
     val = rcg_config_dict.get(rcg_config_name, None)
     if not val:
         raise Exception("Could not get %s config for rcg %s" %

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/testdata/cluster/.gitignore
----------------------------------------------------------------------
diff --git a/testdata/cluster/.gitignore b/testdata/cluster/.gitignore
index c362112..a8aeb3e 100644
--- a/testdata/cluster/.gitignore
+++ b/testdata/cluster/.gitignore
@@ -1,2 +1 @@
-/cdh*
 /kudu

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/testdata/cluster/admin
----------------------------------------------------------------------
diff --git a/testdata/cluster/admin b/testdata/cluster/admin
index 98969ca..bbb4755 100755
--- a/testdata/cluster/admin
+++ b/testdata/cluster/admin
@@ -20,13 +20,12 @@
 
 # This will create/control/destroy a local hdfs+yarn cluster.
 #
-# The original idea was to run each node on a different loopback address but
-# https://jira.cloudera.com/browse/CDH-16602 makes that impossible for now. So 
all roles
-# run on 127.0.0.1, just like the standard mini cluster included with hadoop. 
The
-# difference is with this cluster, each role runs in its own process and has 
its own
+# All roles run on 127.0.0.1, just like the standard mini cluster included 
with hadoop.
+# The difference is with this cluster, each role runs in its own process and 
has its own
 # configs. For each node, the layout of the configs, logs, start/stop scripts, 
etc, is
 # kept as close as possible to a real cluster. For example, the first node 
will live in
 # the dir "cdh-<version>/node-1" and its logs would be at 
"cdh-<version>/node-1/var/log".
+# TODO: Run each node on its own IP address, e.g. 127.0.0.1, 127.0.0.2, and so 
on.
 
 set -euo pipefail
 trap 'echo Error in $0 at line $LINENO: $(awk "NR == $LINENO" $0)' ERR

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/testdata/datasets/functional/schema_constraints.csv
----------------------------------------------------------------------
diff --git a/testdata/datasets/functional/schema_constraints.csv 
b/testdata/datasets/functional/schema_constraints.csv
index a22a432..4c9da6d 100644
--- a/testdata/datasets/functional/schema_constraints.csv
+++ b/testdata/datasets/functional/schema_constraints.csv
@@ -123,7 +123,7 @@ table_name:widerow, constraint:exclude, 
table_format:hbase/none/none
 table_name:nullformat_custom, constraint:exclude, table_format:hbase/none/none
 table_name:unsupported_types, constraint:exclude, table_format:hbase/none/none
 
-# On CDH4, decimal can only be tested on formats Impala can write to (text and 
parquet)
+# Decimal can only be tested on formats Impala can write to (text and parquet).
 # TODO: add Avro once Hive or Impala can write Avro decimals
 table_name:decimal_tbl, constraint:restrict_to, table_format:text/none/none
 table_name:decimal_tiny, constraint:restrict_to, table_format:text/none/none

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/testdata/workloads/functional-planner/queries/PlannerTest/hdfs.test
----------------------------------------------------------------------
diff --git 
a/testdata/workloads/functional-planner/queries/PlannerTest/hdfs.test 
b/testdata/workloads/functional-planner/queries/PlannerTest/hdfs.test
index 880017b..785d111 100644
--- a/testdata/workloads/functional-planner/queries/PlannerTest/hdfs.test
+++ b/testdata/workloads/functional-planner/queries/PlannerTest/hdfs.test
@@ -623,7 +623,7 @@ NODE 0:
   HDFS SPLIT 
hdfs://localhost:20500/test-warehouse/alltypesagg/year=2010/month=1/day=8/100108.txt
 0:76263
   HDFS SPLIT 
hdfs://localhost:20500/test-warehouse/alltypesagg/year=2010/month=1/day=__HIVE_DEFAULT_PARTITION__/000000_0
 0:72759
 ====
-# Predicates on a partition key with no values (see CDH-20089)
+# Predicates on a partition key with no values (see IMPALA-4128).
 select * from functional.emptytable where f2 = 10
 ---- PLAN
 00:SCAN HDFS [functional.emptytable]

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/testdata/workloads/functional-planner/queries/PlannerTest/lineage.test
----------------------------------------------------------------------
diff --git 
a/testdata/workloads/functional-planner/queries/PlannerTest/lineage.test 
b/testdata/workloads/functional-planner/queries/PlannerTest/lineage.test
index 9ba2ab7..20c117e 100644
--- a/testdata/workloads/functional-planner/queries/PlannerTest/lineage.test
+++ b/testdata/workloads/functional-planner/queries/PlannerTest/lineage.test
@@ -3435,7 +3435,7 @@ from
     ]
 }
 ====
-# Create view statement (CDH-26149)
+# Create view statement (IMPALA-4219)
 create view test_view_lineage as select id from functional.alltypestiny
 ---- LINEAGE
 {
@@ -3468,7 +3468,7 @@ create view test_view_lineage as select id from 
functional.alltypestiny
     ]
 }
 ====
-# Create view with nested inline views (CDH-26149)
+# Create view with nested inline views (IMPALA-4219)
 create view test_view_lineage (a, b) as select c1, x2
 from (
   select c1, min(c2) x2
@@ -3547,7 +3547,7 @@ limit 0
     ]
 }
 ====
-# Create view with complex select statement (CDH-26149)
+# Create view with complex select statement (IMPALA-4219)
 create view test_view_lineage (a1, a2, a3, a4, a5, a6, a7) as
   select x.smallint_col, x.id, x.tinyint_col, c.id, x.int_col, x.float_col, 
c.string_col
   from functional.alltypessmall c
@@ -3734,7 +3734,7 @@ create view test_view_lineage (a1, a2, a3, a4, a5, a6, 
a7) as
     ]
 }
 ====
-# Create view with select statement containing an analytic function (CDH-26149)
+# Create view with select statement containing an analytic function 
(IMPALA-4219)
 create view test_view_lineage as
   select * from (
     select sum(a.tinyint_col) over (partition by a.smallint_col order by a.id),
@@ -3878,7 +3878,7 @@ create view test_view_lineage as
     ]
 }
 ====
-# Alter view statement (CDH-26149)
+# Alter view statement (IMPALA-4219)
 alter view functional.alltypes_view as select id from functional.alltypestiny
 ---- LINEAGE
 {

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/testdata/workloads/functional-query/queries/QueryTest/hbase-scan-node.test
----------------------------------------------------------------------
diff --git 
a/testdata/workloads/functional-query/queries/QueryTest/hbase-scan-node.test 
b/testdata/workloads/functional-query/queries/QueryTest/hbase-scan-node.test
index f23fa56..55415eb 100644
--- a/testdata/workloads/functional-query/queries/QueryTest/hbase-scan-node.test
+++ b/testdata/workloads/functional-query/queries/QueryTest/hbase-scan-node.test
@@ -636,7 +636,7 @@ select * from alltypessmallbinary
 INT, BIGINT, BOOLEAN, STRING, DOUBLE, FLOAT, INT, INT, SMALLINT, STRING, 
TIMESTAMP, TINYINT, INT
 ====
 ---- QUERY
-# Scan an HBase table with multiple column families (CDH-18969)
+# Scan an HBase table with multiple column families (IMPALA-4220)
 select * from hbasecolumnfamilies
 ---- RESULTS
 0,true,0,0,0,0,'01/01/09',0,0,'0',2009-01-01 00:00:00
@@ -649,4 +649,4 @@ select * from hbasecolumnfamilies
 7,false,1,1,1,10,'04/01/09',10.1,1.100000023841858,'1',2009-04-01 00:01:00
 ---- TYPES
 INT, BOOLEAN, TINYINT, SMALLINT, INT, BIGINT, STRING, DOUBLE, FLOAT, STRING, 
TIMESTAMP
-====
\ No newline at end of file
+====

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/tests/comparison/cluster.py
----------------------------------------------------------------------
diff --git a/tests/comparison/cluster.py b/tests/comparison/cluster.py
index bfb8fa4..ab1e4f3 100644
--- a/tests/comparison/cluster.py
+++ b/tests/comparison/cluster.py
@@ -630,7 +630,7 @@ class Impalad(object):
     try:
       self._request_web_page("/cancel_query", params={"query_id": id})
     except requests.exceptions.HTTPError as e:
-      # XXX: Handle losing the race
+      # TODO: Handle losing the race
       raise e
 
   def shell(self, cmd, timeout_secs=DEFAULT_TIMEOUT):

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/tests/comparison/discrepancy_searcher.py
----------------------------------------------------------------------
diff --git a/tests/comparison/discrepancy_searcher.py 
b/tests/comparison/discrepancy_searcher.py
index a1e60e2..57e284a 100755
--- a/tests/comparison/discrepancy_searcher.py
+++ b/tests/comparison/discrepancy_searcher.py
@@ -247,7 +247,7 @@ class QueryResultComparator(object):
 class QueryExecutor(object):
   '''Concurrently executes queries'''
 
-  # XXX: Set to false while IMPALA-3336 is a problem. Disabling random query 
options
+  # TODO: Set to false while IMPALA-3336 is a problem. Disabling random query 
options
   # seems to reduce IMPALA-3336 occurances.
   ENABLE_RANDOM_QUERY_OPTIONS = False
 
@@ -379,9 +379,9 @@ class QueryExecutor(object):
           cursor.conn.kill()
           LOG.debug('Kill connection')
         try:
-          # XXX: Sometimes this takes a very long time causing the program to 
appear to
-          #      hang. Maybe this should be done in another thread so a 
timeout can be
-          #      applied?
+          # TODO: Sometimes this takes a very long time causing the program to 
appear to
+          #       hang. Maybe this should be done in another thread so a 
timeout can be
+          #       applied?
           cursor.close()
         except Exception as e:
           LOG.info('Error closing cursor: %s', e)

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/tests/comparison/leopard/impala_docker_env.py
----------------------------------------------------------------------
diff --git a/tests/comparison/leopard/impala_docker_env.py 
b/tests/comparison/leopard/impala_docker_env.py
index 3f1d292..715ad65 100755
--- a/tests/comparison/leopard/impala_docker_env.py
+++ b/tests/comparison/leopard/impala_docker_env.py
@@ -32,7 +32,7 @@ import os
 IMPALA_HOME = '/home/dev/Impala'
 CORE_PATH = '/tmp/core_files'
 DEFAULT_BRANCH_NAME = 'origin/cdh5-trunk'
-DEFAULT_DOCKER_IMAGE_NAME = 
'impala-desktop.ca.cloudera.com:5000/ubuntu-14.04:cdh5-trunk'
+DEFAULT_DOCKER_IMAGE_NAME = 'cloudera/impala-dev'
 DOCKER_USER_NAME = 'dev'
 NUM_START_ATTEMPTS = 50
 NUM_FABRIC_ATTEMPTS = 50
@@ -152,10 +152,11 @@ class ImpalaDockerEnv(object):
     elif SHOULD_BUILD_IMPALA:
       build_command = (
           'docker-boot && cd {IMPALA_HOME} '
-          '&& git fetch --all && git checkout origin/cdh5-trunk '
+          '&& git fetch --all && git checkout DEFAULT_BRANCH_NAME '
           '&& source {IMPALA_HOME}/bin/impala-config.sh '
           '&& {IMPALA_HOME}/buildall.sh -notests').format(
               IMPALA_HOME = IMPALA_HOME,
+              DEFAULT_BRANCH_NAME = DEFAULT_BRANCH_NAME,
               CORE_PATH = CORE_PATH)
 
     if build_command:

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/tests/comparison/query.py
----------------------------------------------------------------------
diff --git a/tests/comparison/query.py b/tests/comparison/query.py
index b928bf3..f6ec65c 100644
--- a/tests/comparison/query.py
+++ b/tests/comparison/query.py
@@ -336,8 +336,8 @@ class Subquery(ValExpr):
      result set.
 
   '''
-  # XXX: So far it seems fine to use this class for both scalar/non scalar 
cases but
-  #      this could lead to unexpected behavior or be a silent cause of 
problems...
+  # TODO: So far it seems fine to use this class for both scalar/non scalar 
cases but
+  #       this could lead to unexpected behavior or be a silent cause of 
problems...
 
   def __init__(self, query):
     self.query = query

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/tests/custom_cluster/test_insert_behaviour.py
----------------------------------------------------------------------
diff --git a/tests/custom_cluster/test_insert_behaviour.py 
b/tests/custom_cluster/test_insert_behaviour.py
index 9cec88f..6df8ed7 100644
--- a/tests/custom_cluster/test_insert_behaviour.py
+++ b/tests/custom_cluster/test_insert_behaviour.py
@@ -120,7 +120,7 @@ class 
TestInsertBehaviourCustomCluster(CustomClusterTestSuite):
       self.execute_query_expect_success(client, "INSERT INTO %s"
                                         " PARTITION(p1=1, p2=3, p3=4) 
VALUES(1)" % TEST_TBL)
       # Would be 777 if inheritance was enabled
-      if not IS_ISILON: # CDH-27688
+      if not IS_ISILON: # IMPALA-4221
         self._check_partition_perms("p1=1/p2=3/", default_perms)
       self._check_partition_perms("p1=1/p2=3/p3=4/", default_perms)
     finally:

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/tests/metadata/test_compute_stats.py
----------------------------------------------------------------------
diff --git a/tests/metadata/test_compute_stats.py 
b/tests/metadata/test_compute_stats.py
index 6f6d31a..9ecdcdf 100644
--- a/tests/metadata/test_compute_stats.py
+++ b/tests/metadata/test_compute_stats.py
@@ -42,7 +42,8 @@ class TestComputeStats(ImpalaTestSuite):
   @SkipIfLocal.hdfs_blocks
   def test_compute_stats(self, vector, unique_database):
     self.run_test_case('QueryTest/compute-stats', vector, unique_database)
-    # Test compute stats on decimal columns separately so we can vary between 
CDH4/5
+    # Test compute stats on decimal columns separately so we can vary between 
platforms
+    # with and without write support for decimals (Hive < 0.11 and >= 0.11).
     self.run_test_case('QueryTest/compute-stats-decimal', vector, 
unique_database)
 
   def test_compute_stats_incremental(self, vector, unique_database):

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/tests/metadata/test_hdfs_encryption.py
----------------------------------------------------------------------
diff --git a/tests/metadata/test_hdfs_encryption.py 
b/tests/metadata/test_hdfs_encryption.py
index 196b2ea..a257de0 100644
--- a/tests/metadata/test_hdfs_encryption.py
+++ b/tests/metadata/test_hdfs_encryption.py
@@ -146,7 +146,7 @@ class TestHdfsEncryption(ImpalaTestSuite):
   def test_drop_partition_encrypt(self):
     """Verifies if alter <tbl> drop partition purge works in case
     where the Trash dir and partition dir are in different encryption
-    zones. Check CDH-31350 for details"""
+    zones. Check IMPALA-2310 for details"""
     self.client.execute("create table {0}.t1(i int) partitioned\
       by (j int)".format(TEST_DB))
     # Add three partitions (j=1), (j=2), (j=3) to table t1

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/tests/query_test/test_decimal_queries.py
----------------------------------------------------------------------
diff --git a/tests/query_test/test_decimal_queries.py 
b/tests/query_test/test_decimal_queries.py
index 0404ea1..f95d940 100644
--- a/tests/query_test/test_decimal_queries.py
+++ b/tests/query_test/test_decimal_queries.py
@@ -35,8 +35,9 @@ class TestDecimalQueries(ImpalaTestSuite):
     cls.TestMatrix.add_dimension(
         TestDimension('batch_size', *TestDecimalQueries.BATCH_SIZES))
 
-    # On CDH4, hive does not support decimal so we can't run these tests 
against
-    # the other file formats. Enable them on C5.
+    # Hive < 0.11 does not support decimal so we can't run these tests against 
the other
+    # file formats.
+    # TODO: Enable them on Hive >= 0.11.
     cls.TestMatrix.add_constraint(lambda v:\
         (v.get_value('table_format').file_format == 'text' and
          v.get_value('table_format').compression_codec == 'none') or

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/tests/shell/test_shell_commandline.py
----------------------------------------------------------------------
diff --git a/tests/shell/test_shell_commandline.py 
b/tests/shell/test_shell_commandline.py
index 31cf570..16a68ae 100644
--- a/tests/shell/test_shell_commandline.py
+++ b/tests/shell/test_shell_commandline.py
@@ -200,7 +200,7 @@ class TestImpalaShell(ImpalaTestSuite):
     assert 'Problem parsing file' in result.stderr
 
   def test_no_warnings_in_log_with_quiet_mode(self):
-    """Regression test for CDH-21036."""
+    """Regression test for IMPALA-4222."""
     args = ('-q "set abort_on_error=false;'
             ' select count(*) from functional_seq_snap.bad_seq_snap" --quiet')
     result = run_impala_shell_cmd(args)

http://git-wip-us.apache.org/repos/asf/incubator-impala/blob/ef4c9958/tests/stress/concurrent_select.py
----------------------------------------------------------------------
diff --git a/tests/stress/concurrent_select.py 
b/tests/stress/concurrent_select.py
index 1d1d142..b76d713 100755
--- a/tests/stress/concurrent_select.py
+++ b/tests/stress/concurrent_select.py
@@ -335,8 +335,8 @@ class StressRunner(object):
        If a query completes without error, the result will be verified. An 
error
        will be raised upon a result mismatch.
     """
-    # XXX: The state from a previous run should be cleared out. This isn't 
really a
-    #      problem now because the one caller (main()) never calls a second 
time.
+    # TODO: The state from a previous run should be cleared out. This isn't 
really a
+    #       problem now because the one caller (main()) never calls a second 
time.
 
     if self.startup_queries_per_sec <= 0:
       raise Exception("Startup queries per second must be positive")


Reply via email to