git commit: PHOENIX-1175 Add setReadOnly and setFetchSize options (Alex Kamil)

2014-08-15 Thread jamestaylor
Repository: phoenix
Updated Branches:
  refs/heads/4.0 80a3c4806 -> db3b65669


PHOENIX-1175 Add setReadOnly and setFetchSize options (Alex Kamil)


Project: http://git-wip-us.apache.org/repos/asf/phoenix/repo
Commit: http://git-wip-us.apache.org/repos/asf/phoenix/commit/db3b6566
Tree: http://git-wip-us.apache.org/repos/asf/phoenix/tree/db3b6566
Diff: http://git-wip-us.apache.org/repos/asf/phoenix/diff/db3b6566

Branch: refs/heads/4.0
Commit: db3b6566937877060a582f4a3bceef9eb32eb04d
Parents: 80a3c48
Author: James Taylor 
Authored: Fri Aug 15 23:56:34 2014 -0700
Committer: James Taylor 
Committed: Fri Aug 15 23:57:33 2014 -0700

--
 .../org/apache/phoenix/end2end/ReadOnlyIT.java  | 98 
 .../phoenix/exception/SQLExceptionCode.java |  7 +-
 .../apache/phoenix/jdbc/PhoenixConnection.java  |  9 +-
 .../apache/phoenix/jdbc/PhoenixStatement.java   | 16 +++-
 4 files changed, 120 insertions(+), 10 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/phoenix/blob/db3b6566/phoenix-core/src/it/java/org/apache/phoenix/end2end/ReadOnlyIT.java
--
diff --git 
a/phoenix-core/src/it/java/org/apache/phoenix/end2end/ReadOnlyIT.java 
b/phoenix-core/src/it/java/org/apache/phoenix/end2end/ReadOnlyIT.java
new file mode 100644
index 000..515acac
--- /dev/null
+++ b/phoenix-core/src/it/java/org/apache/phoenix/end2end/ReadOnlyIT.java
@@ -0,0 +1,98 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.phoenix.end2end;
+
+import static org.apache.phoenix.util.TestUtil.TEST_PROPERTIES;
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertFalse;
+import static org.junit.Assert.assertTrue;
+import static org.junit.Assert.fail;
+
+import java.sql.Connection;
+import java.sql.DriverManager;
+import java.sql.PreparedStatement;
+import java.sql.ResultSet;
+import java.sql.SQLException;
+import java.util.Properties;
+import java.sql.ResultSetMetaData;
+
+import org.apache.phoenix.util.PropertiesUtil;
+import org.junit.Test;
+import org.junit.experimental.categories.Category;
+
+@Category(HBaseManagedTimeTest.class)
+public class ReadOnlyIT extends BaseHBaseManagedTimeIT {
+
+@Test
+public void testConnectionReadOnly() throws Exception {
+
+Properties props = PropertiesUtil.deepCopy(TEST_PROPERTIES);
+Connection conn = DriverManager.getConnection(getUrl(), props);
+String ddl = "CREATE TABLE test_table " +
+"  (row varchar not null, col1 integer" +
+"  CONSTRAINT pk PRIMARY KEY (row))\n"; 
+createTestTable(getUrl(), ddl);
+
+String query = "UPSERT INTO test_table(row, col1) VALUES('row1', 777)";
+PreparedStatement statement = conn.prepareStatement(query);
+statement.executeUpdate();
+conn.commit();
+
+   try{
+   conn.setReadOnly(true);
+assertTrue(conn.isReadOnly());
+   ddl = "CREATE TABLE test_table2 " +
+   "  (row varchar not null, col1 integer" +
+   "  CONSTRAINT pk PRIMARY KEY (row))\n";
+   statement = conn.prepareStatement(ddl);
+   statement.executeUpdate();
+   conn.commit();
+   fail();
+   } catch (SQLException e) {
+  assertTrue(e.getMessage(), e.getMessage().contains("ERROR 518 
(25502): Mutations are not permitted for a read-only connection."));
+}
+ 
+   try {  
+query = "UPSERT INTO test_table(row, col1) VALUES('row1', 
888)";
+statement = conn.prepareStatement(query);
+statement.executeUpdate();
+conn.commit();
+fail();
+} catch (SQLException e) {
+  assertTrue(e.getMessage(), e.getMessage().contains("ERROR 518 
(25502): Mutations are not permitted for a read-only connection."));
+}
+
+   conn.setReadOnly(false);
+assertFalse(conn.isReadOnly());
+ddl = 

git commit: PHOENIX-1175 Add setReadOnly and setFetchSize options (Alex Kamil)

2014-08-15 Thread jamestaylor
Repository: phoenix
Updated Branches:
  refs/heads/master ebb6a7adb -> 355369ad5


PHOENIX-1175 Add setReadOnly and setFetchSize options (Alex Kamil)


Project: http://git-wip-us.apache.org/repos/asf/phoenix/repo
Commit: http://git-wip-us.apache.org/repos/asf/phoenix/commit/355369ad
Tree: http://git-wip-us.apache.org/repos/asf/phoenix/tree/355369ad
Diff: http://git-wip-us.apache.org/repos/asf/phoenix/diff/355369ad

Branch: refs/heads/master
Commit: 355369ad5bd9a61d1b19e6bebec32ea8cb8b8a0c
Parents: ebb6a7a
Author: James Taylor 
Authored: Fri Aug 15 23:56:34 2014 -0700
Committer: James Taylor 
Committed: Fri Aug 15 23:56:34 2014 -0700

--
 .../org/apache/phoenix/end2end/ReadOnlyIT.java  | 98 
 .../phoenix/exception/SQLExceptionCode.java |  7 +-
 .../apache/phoenix/jdbc/PhoenixConnection.java  |  9 +-
 .../apache/phoenix/jdbc/PhoenixStatement.java   | 16 +++-
 4 files changed, 120 insertions(+), 10 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/phoenix/blob/355369ad/phoenix-core/src/it/java/org/apache/phoenix/end2end/ReadOnlyIT.java
--
diff --git 
a/phoenix-core/src/it/java/org/apache/phoenix/end2end/ReadOnlyIT.java 
b/phoenix-core/src/it/java/org/apache/phoenix/end2end/ReadOnlyIT.java
new file mode 100644
index 000..515acac
--- /dev/null
+++ b/phoenix-core/src/it/java/org/apache/phoenix/end2end/ReadOnlyIT.java
@@ -0,0 +1,98 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.phoenix.end2end;
+
+import static org.apache.phoenix.util.TestUtil.TEST_PROPERTIES;
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertFalse;
+import static org.junit.Assert.assertTrue;
+import static org.junit.Assert.fail;
+
+import java.sql.Connection;
+import java.sql.DriverManager;
+import java.sql.PreparedStatement;
+import java.sql.ResultSet;
+import java.sql.SQLException;
+import java.util.Properties;
+import java.sql.ResultSetMetaData;
+
+import org.apache.phoenix.util.PropertiesUtil;
+import org.junit.Test;
+import org.junit.experimental.categories.Category;
+
+@Category(HBaseManagedTimeTest.class)
+public class ReadOnlyIT extends BaseHBaseManagedTimeIT {
+
+@Test
+public void testConnectionReadOnly() throws Exception {
+
+Properties props = PropertiesUtil.deepCopy(TEST_PROPERTIES);
+Connection conn = DriverManager.getConnection(getUrl(), props);
+String ddl = "CREATE TABLE test_table " +
+"  (row varchar not null, col1 integer" +
+"  CONSTRAINT pk PRIMARY KEY (row))\n"; 
+createTestTable(getUrl(), ddl);
+
+String query = "UPSERT INTO test_table(row, col1) VALUES('row1', 777)";
+PreparedStatement statement = conn.prepareStatement(query);
+statement.executeUpdate();
+conn.commit();
+
+   try{
+   conn.setReadOnly(true);
+assertTrue(conn.isReadOnly());
+   ddl = "CREATE TABLE test_table2 " +
+   "  (row varchar not null, col1 integer" +
+   "  CONSTRAINT pk PRIMARY KEY (row))\n";
+   statement = conn.prepareStatement(ddl);
+   statement.executeUpdate();
+   conn.commit();
+   fail();
+   } catch (SQLException e) {
+  assertTrue(e.getMessage(), e.getMessage().contains("ERROR 518 
(25502): Mutations are not permitted for a read-only connection."));
+}
+ 
+   try {  
+query = "UPSERT INTO test_table(row, col1) VALUES('row1', 
888)";
+statement = conn.prepareStatement(query);
+statement.executeUpdate();
+conn.commit();
+fail();
+} catch (SQLException e) {
+  assertTrue(e.getMessage(), e.getMessage().contains("ERROR 518 
(25502): Mutations are not permitted for a read-only connection."));
+}
+
+   conn.setReadOnly(false);
+assertFalse(conn.isReadOnly());
+

Apache-Phoenix | Master | Hadoop1 | Build Successful

2014-08-15 Thread Apache Jenkins Server
Master branch build status Successful
Source repository https://git-wip-us.apache.org/repos/asf/incubator-phoenix.git

Last Successful Compiled Artifacts https://builds.apache.org/job/Phoenix-master-hadoop1/lastSuccessfulBuild/artifact/

Last Complete Test Report https://builds.apache.org/job/Phoenix-master-hadoop1/lastCompletedBuild/testReport/

Changes
[jamestaylor] PHOENIX-1174 Rename and move properties using existing convention

[jeffreyz] PHOENIX-1173: MutableIndexFailureIT.java doesn't finish sometimes or is flappy.



Jenkins build is back to normal : Phoenix | 3.0 | Hadoop1 #187

2014-08-15 Thread Apache Jenkins Server
See 



Build failed in Jenkins: Phoenix | 4.0 | Hadoop1 #279

2014-08-15 Thread Apache Jenkins Server
See 

Changes:

[jamestaylor] PHOENIX-1174 Rename and move properties using existing convention

[jeffreyz] PHOENIX-1173: MutableIndexFailureIT.java doesn't finish sometimes or 
is flappy.

--
[...truncated 459 lines...]
Running org.apache.phoenix.end2end.UpsertSelectIT
Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 6.033 sec - in 
org.apache.phoenix.end2end.ToCharFunctionIT
Running org.apache.phoenix.end2end.VariableLengthPKIT
Tests run: 16, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 21.375 sec - 
in org.apache.phoenix.end2end.DerivedTableIT
Running org.apache.phoenix.end2end.MultiCfQueryExecIT
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 25.723 sec - in 
org.apache.phoenix.end2end.ColumnProjectionOptimizationIT
Running org.apache.phoenix.end2end.CoalesceFunctionIT
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 11.728 sec - in 
org.apache.phoenix.end2end.CoalesceFunctionIT
Running org.apache.phoenix.end2end.CreateTableIT
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 20.248 sec - in 
org.apache.phoenix.end2end.MultiCfQueryExecIT
Running org.apache.phoenix.end2end.IsNullIT
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 9.81 sec - in 
org.apache.phoenix.end2end.IsNullIT
Running org.apache.phoenix.end2end.StddevIT
Tests run: 77, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 62.213 sec - 
in org.apache.phoenix.end2end.NotQueryIT
Running org.apache.phoenix.end2end.RowValueConstructorIT
Tests run: 49, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 39.441 sec - 
in org.apache.phoenix.end2end.VariableLengthPKIT
Running org.apache.phoenix.end2end.ArrayIT
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 6.907 sec - in 
org.apache.phoenix.end2end.StddevIT
Running org.apache.phoenix.end2end.GroupByCaseIT
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.415 sec - in 
org.apache.phoenix.end2end.GroupByCaseIT
Running org.apache.phoenix.end2end.SpooledOrderByIT
Tests run: 63, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 55.651 sec - 
in org.apache.phoenix.end2end.CaseStatementIT
Running org.apache.phoenix.end2end.ClientTimeArithmeticQueryIT
Tests run: 3, Failures: 0, Errors: 0, Skipped: 1, Time elapsed: 9.247 sec - in 
org.apache.phoenix.end2end.SpooledOrderByIT
Running org.apache.phoenix.end2end.GroupByIT
Tests run: 12, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 63.682 sec - 
in org.apache.phoenix.end2end.UpsertSelectIT
Running org.apache.phoenix.end2end.ProductMetricsIT
Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 49.8 sec - in 
org.apache.phoenix.end2end.CreateTableIT
Tests run: 48, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 34.586 sec - 
in org.apache.phoenix.end2end.ArrayIT
Tests run: 33, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 38.708 sec - 
in org.apache.phoenix.end2end.RowValueConstructorIT
Tests run: 61, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 32.933 sec - 
in org.apache.phoenix.end2end.ProductMetricsIT
Tests run: 91, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 75.776 sec - 
in org.apache.phoenix.end2end.GroupByIT
Tests run: 182, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 155.872 sec - 
in org.apache.phoenix.end2end.QueryIT
Tests run: 203, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 136.341 sec - 
in org.apache.phoenix.end2end.ClientTimeArithmeticQueryIT

Results :

Tests run: 1268, Failures: 0, Errors: 0, Skipped: 3

[INFO] 
[INFO] --- maven-failsafe-plugin:2.17:integration-test (HBaseManagedTimeTests) 
@ phoenix-core ---
[INFO] Failsafe report directory: 
/x1/jenkins/jenkins-slave/workspace/Phoenix-4.0-hadoop1/phoenix-core/target/failsafe-reports
[INFO] parallel='none', perCoreThreadCount=true, threadCount=0, 
useUnlimitedThreads=false, threadCountSuites=0, threadCountClasses=0, 
threadCountMethods=0, parallelOptimized=true

---
 T E S T S
---

---
 T E S T S
---
Running org.apache.phoenix.end2end.QueryExecWithoutSCNIT
Running org.apache.phoenix.end2end.TenantSpecificViewIndexIT
Running org.apache.phoenix.end2end.PhoenixEncodeDecodeIT
Running org.apache.phoenix.end2end.DeleteIT
Running org.apache.phoenix.end2end.BinaryRowKeyIT
Running org.apache.phoenix.end2end.TenantSpecificViewIndexSaltedIT
Running org.apache.phoenix.end2end.SkipScanQueryIT
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.229 sec - in 
org.apache.phoenix.end2end.BinaryRowKeyIT
Running org.apache.phoenix.end2end.EncodeFunctionIT
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 9.065 sec - in 
org.apache.phoenix.end2end.QueryExecWithoutSCNIT
Running org.apache.phoenix.end2end.TimezoneOffsetFunctionIT

git commit: PHOENIX-1173: MutableIndexFailureIT.java doesn't finish sometimes or is flappy.

2014-08-15 Thread jeffreyz
Repository: phoenix
Updated Branches:
  refs/heads/master 367662dc8 -> ebb6a7adb


PHOENIX-1173: MutableIndexFailureIT.java doesn't finish sometimes or is flappy.


Project: http://git-wip-us.apache.org/repos/asf/phoenix/repo
Commit: http://git-wip-us.apache.org/repos/asf/phoenix/commit/ebb6a7ad
Tree: http://git-wip-us.apache.org/repos/asf/phoenix/tree/ebb6a7ad
Diff: http://git-wip-us.apache.org/repos/asf/phoenix/diff/ebb6a7ad

Branch: refs/heads/master
Commit: ebb6a7adb9134eb2413950796bd5f4e80a250e7d
Parents: 367662d
Author: Jeffrey Zhong 
Authored: Fri Aug 15 14:02:51 2014 -0700
Committer: Jeffrey Zhong 
Committed: Fri Aug 15 16:05:55 2014 -0700

--
 .../apache/phoenix/coprocessor/MetaDataEndpointImpl.java| 2 ++
 .../apache/phoenix/coprocessor/MetaDataRegionObserver.java  | 9 ++---
 2 files changed, 8 insertions(+), 3 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/phoenix/blob/ebb6a7ad/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
--
diff --git 
a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
 
b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
index b99483b..5b43a90 100644
--- 
a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
+++ 
b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
@@ -1508,6 +1508,8 @@ public class MetaDataEndpointImpl extends 
MetaDataProtocol implements Coprocesso
 dataTableKey = SchemaUtil.getTableKey(tenantId, 
schemaName, dataTableKV.getValue());
 }
 if(dataTableKey != null) {
+// make a copy of tableMetadata
+tableMetadata = new ArrayList(tableMetadata);
 // insert an empty KV to trigger time stamp update on 
data table row
 Put p = new Put(dataTableKey);
 p.add(TABLE_FAMILY_BYTES, 
QueryConstants.EMPTY_COLUMN_BYTES, timeStamp, ByteUtil.EMPTY_BYTE_ARRAY);

http://git-wip-us.apache.org/repos/asf/phoenix/blob/ebb6a7ad/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataRegionObserver.java
--
diff --git 
a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataRegionObserver.java
 
b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataRegionObserver.java
index 6ce0148..822ced8 100644
--- 
a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataRegionObserver.java
+++ 
b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataRegionObserver.java
@@ -28,6 +28,9 @@ import java.util.ArrayList;
 import java.util.List;
 import java.util.Timer;
 import java.util.TimerTask;
+import java.util.concurrent.ScheduledFuture;
+import java.util.concurrent.ScheduledThreadPoolExecutor;
+import java.util.concurrent.TimeUnit;
 import java.util.concurrent.atomic.AtomicInteger;
 
 import org.apache.commons.logging.Log;
@@ -65,14 +68,14 @@ import org.apache.phoenix.util.SchemaUtil;
  */
 public class MetaDataRegionObserver extends BaseRegionObserver {
 public static final Log LOG = 
LogFactory.getLog(MetaDataRegionObserver.class);
-protected Timer scheduleTimer = new Timer(true);
+protected ScheduledThreadPoolExecutor executor = new 
ScheduledThreadPoolExecutor(1);
 private boolean enableRebuildIndex = 
QueryServicesOptions.DEFAULT_INDEX_FAILURE_HANDLING_REBUILD;
 private long rebuildIndexTimeInterval = 
QueryServicesOptions.DEFAULT_INDEX_FAILURE_HANDLING_REBUILD_INTERVAL;
   
 @Override
 public void preClose(final ObserverContext c,
 boolean abortRequested) {
-scheduleTimer.cancel();
+executor.shutdownNow();
 
GlobalCache.getInstance(c.getEnvironment()).getMetaDataCache().invalidateAll();
 }
 
@@ -113,7 +116,7 @@ public class MetaDataRegionObserver extends 
BaseRegionObserver {
 // starts index rebuild schedule work
 BuildIndexScheduleTask task = new 
BuildIndexScheduleTask(e.getEnvironment());
 // run scheduled task every 10 secs
-scheduleTimer.schedule(task, 1, rebuildIndexTimeInterval);
+executor.scheduleAtFixedRate(task, 1, 
rebuildIndexTimeInterval, TimeUnit.MILLISECONDS);
 } catch (ClassNotFoundException ex) {
 LOG.error("BuildIndexScheduleTask cannot start!", ex);
 }



git commit: PHOENIX-1173: MutableIndexFailureIT.java doesn't finish sometimes or is flappy.

2014-08-15 Thread jeffreyz
Repository: phoenix
Updated Branches:
  refs/heads/4.0 5aa381516 -> 80a3c4806


PHOENIX-1173: MutableIndexFailureIT.java doesn't finish sometimes or is flappy.


Project: http://git-wip-us.apache.org/repos/asf/phoenix/repo
Commit: http://git-wip-us.apache.org/repos/asf/phoenix/commit/80a3c480
Tree: http://git-wip-us.apache.org/repos/asf/phoenix/tree/80a3c480
Diff: http://git-wip-us.apache.org/repos/asf/phoenix/diff/80a3c480

Branch: refs/heads/4.0
Commit: 80a3c4806534594e94fdc204b37a7103a7eb787e
Parents: 5aa3815
Author: Jeffrey Zhong 
Authored: Fri Aug 15 14:02:51 2014 -0700
Committer: Jeffrey Zhong 
Committed: Fri Aug 15 14:50:29 2014 -0700

--
 .../apache/phoenix/coprocessor/MetaDataEndpointImpl.java| 2 ++
 .../apache/phoenix/coprocessor/MetaDataRegionObserver.java  | 9 ++---
 2 files changed, 8 insertions(+), 3 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/phoenix/blob/80a3c480/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
--
diff --git 
a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
 
b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
index b99483b..5b43a90 100644
--- 
a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
+++ 
b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
@@ -1508,6 +1508,8 @@ public class MetaDataEndpointImpl extends 
MetaDataProtocol implements Coprocesso
 dataTableKey = SchemaUtil.getTableKey(tenantId, 
schemaName, dataTableKV.getValue());
 }
 if(dataTableKey != null) {
+// make a copy of tableMetadata
+tableMetadata = new ArrayList(tableMetadata);
 // insert an empty KV to trigger time stamp update on 
data table row
 Put p = new Put(dataTableKey);
 p.add(TABLE_FAMILY_BYTES, 
QueryConstants.EMPTY_COLUMN_BYTES, timeStamp, ByteUtil.EMPTY_BYTE_ARRAY);

http://git-wip-us.apache.org/repos/asf/phoenix/blob/80a3c480/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataRegionObserver.java
--
diff --git 
a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataRegionObserver.java
 
b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataRegionObserver.java
index 6ce0148..822ced8 100644
--- 
a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataRegionObserver.java
+++ 
b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataRegionObserver.java
@@ -28,6 +28,9 @@ import java.util.ArrayList;
 import java.util.List;
 import java.util.Timer;
 import java.util.TimerTask;
+import java.util.concurrent.ScheduledFuture;
+import java.util.concurrent.ScheduledThreadPoolExecutor;
+import java.util.concurrent.TimeUnit;
 import java.util.concurrent.atomic.AtomicInteger;
 
 import org.apache.commons.logging.Log;
@@ -65,14 +68,14 @@ import org.apache.phoenix.util.SchemaUtil;
  */
 public class MetaDataRegionObserver extends BaseRegionObserver {
 public static final Log LOG = 
LogFactory.getLog(MetaDataRegionObserver.class);
-protected Timer scheduleTimer = new Timer(true);
+protected ScheduledThreadPoolExecutor executor = new 
ScheduledThreadPoolExecutor(1);
 private boolean enableRebuildIndex = 
QueryServicesOptions.DEFAULT_INDEX_FAILURE_HANDLING_REBUILD;
 private long rebuildIndexTimeInterval = 
QueryServicesOptions.DEFAULT_INDEX_FAILURE_HANDLING_REBUILD_INTERVAL;
   
 @Override
 public void preClose(final ObserverContext c,
 boolean abortRequested) {
-scheduleTimer.cancel();
+executor.shutdownNow();
 
GlobalCache.getInstance(c.getEnvironment()).getMetaDataCache().invalidateAll();
 }
 
@@ -113,7 +116,7 @@ public class MetaDataRegionObserver extends 
BaseRegionObserver {
 // starts index rebuild schedule work
 BuildIndexScheduleTask task = new 
BuildIndexScheduleTask(e.getEnvironment());
 // run scheduled task every 10 secs
-scheduleTimer.schedule(task, 1, rebuildIndexTimeInterval);
+executor.scheduleAtFixedRate(task, 1, 
rebuildIndexTimeInterval, TimeUnit.MILLISECONDS);
 } catch (ClassNotFoundException ex) {
 LOG.error("BuildIndexScheduleTask cannot start!", ex);
 }



git commit: PHOENIX-1174 Rename and move properties using existing convention

2014-08-15 Thread jamestaylor
Repository: phoenix
Updated Branches:
  refs/heads/master dfcde1046 -> 367662dc8


PHOENIX-1174 Rename and move properties using existing convention


Project: http://git-wip-us.apache.org/repos/asf/phoenix/repo
Commit: http://git-wip-us.apache.org/repos/asf/phoenix/commit/367662dc
Tree: http://git-wip-us.apache.org/repos/asf/phoenix/tree/367662dc
Diff: http://git-wip-us.apache.org/repos/asf/phoenix/diff/367662dc

Branch: refs/heads/master
Commit: 367662dc884433cd3e626b65e4417716966062fb
Parents: dfcde10
Author: James Taylor 
Authored: Fri Aug 15 14:32:14 2014 -0700
Committer: James Taylor 
Committed: Fri Aug 15 14:36:49 2014 -0700

--
 .../phoenix/end2end/index/IndexHandlerIT.java   |  4 ++--
 .../ipc/PhoenixIndexRpcSchedulerFactory.java| 23 +---
 .../org/apache/phoenix/query/QueryServices.java |  9 
 .../phoenix/query/QueryServicesOptions.java | 11 --
 .../org/apache/phoenix/trace/util/Tracing.java  | 18 ++-
 .../PhoenixIndexRpcSchedulerFactoryTest.java|  5 +++--
 6 files changed, 35 insertions(+), 35 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/phoenix/blob/367662dc/phoenix-core/src/it/java/org/apache/phoenix/end2end/index/IndexHandlerIT.java
--
diff --git 
a/phoenix-core/src/it/java/org/apache/phoenix/end2end/index/IndexHandlerIT.java 
b/phoenix-core/src/it/java/org/apache/phoenix/end2end/index/IndexHandlerIT.java
index 8536652..1507d6b 100644
--- 
a/phoenix-core/src/it/java/org/apache/phoenix/end2end/index/IndexHandlerIT.java
+++ 
b/phoenix-core/src/it/java/org/apache/phoenix/end2end/index/IndexHandlerIT.java
@@ -38,7 +38,7 @@ import org.apache.hadoop.hbase.ipc.RpcControllerFactory;
 import org.apache.hadoop.hbase.util.Bytes;
 import org.apache.phoenix.hbase.index.IndexQosRpcControllerFactory;
 import org.apache.phoenix.hbase.index.TableName;
-import org.apache.phoenix.hbase.index.ipc.PhoenixIndexRpcSchedulerFactory;
+import org.apache.phoenix.query.QueryServicesOptions;
 import org.junit.After;
 import org.junit.AfterClass;
 import org.junit.Before;
@@ -159,7 +159,7 @@ public class IndexHandlerIT {
 // check the counts on the rpc controller
 assertEquals("Didn't get the expected number of index priority 
writes!", 1,
 (int) CountingIndexClientRpcController.priorityCounts
-
.get(PhoenixIndexRpcSchedulerFactory.DEFAULT_INDEX_MIN_PRIORITY));
+.get(QueryServicesOptions.DEFAULT_INDEX_MIN_PRIORITY));
 
 table.close();
 }

http://git-wip-us.apache.org/repos/asf/phoenix/blob/367662dc/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/ipc/PhoenixIndexRpcSchedulerFactory.java
--
diff --git 
a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/ipc/PhoenixIndexRpcSchedulerFactory.java
 
b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/ipc/PhoenixIndexRpcSchedulerFactory.java
index 500db7c..8e0b86f 100644
--- 
a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/ipc/PhoenixIndexRpcSchedulerFactory.java
+++ 
b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/ipc/PhoenixIndexRpcSchedulerFactory.java
@@ -26,6 +26,8 @@ import org.apache.hadoop.hbase.ipc.RpcScheduler;
 import org.apache.hadoop.hbase.regionserver.RegionServerServices;
 import org.apache.hadoop.hbase.regionserver.RpcSchedulerFactory;
 import org.apache.hadoop.hbase.regionserver.SimpleRpcSchedulerFactory;
+import org.apache.phoenix.query.QueryServices;
+import org.apache.phoenix.query.QueryServicesOptions;
 
 import com.google.common.base.Preconditions;
 
@@ -37,21 +39,6 @@ public class PhoenixIndexRpcSchedulerFactory implements 
RpcSchedulerFactory {
 
 private static final Log LOG = 
LogFactory.getLog(PhoenixIndexRpcSchedulerFactory.class);
 
-private static final String INDEX_HANDLER_COUNT_KEY =
-"org.apache.phoenix.regionserver.index.handler.count";
-private static final int DEFAULT_INDEX_HANDLER_COUNT = 30;
-
-/**
- * HConstants#HIGH_QOS is the max we will see to a standard table. We go 
higher to differentiate
- * and give some room for things in the middle
- */
-public static final int DEFAULT_INDEX_MIN_PRIORITY = 1000;
-public static final int DEFAULT_INDEX_MAX_PRIORITY = 1050;
-public static final String MIN_INDEX_PRIOIRTY_KEY =
-"org.apache.phoenix.regionserver.index.priority.min";
-public static final String MAX_INDEX_PRIOIRTY_KEY =
-"org.apache.phoenix.regionserver.index.priority.max";
-
 private static final String VERSION_TOO_OLD_FOR_INDEX_RPC =
 "Running an older version of HBase (less than 0.98.4), Phoenix 
index RPC handling cannot be enabled.";
 
@@ -75,9 +62,9 @@ public class PhoenixIndexR

git commit: PHOENIX-1173: MutableIndexFailureIT.java doesn't finish sometimes or is flappy.

2014-08-15 Thread jeffreyz
Repository: phoenix
Updated Branches:
  refs/heads/3.0 19dc23aa5 -> 71cc23c8f


PHOENIX-1173: MutableIndexFailureIT.java doesn't finish sometimes or is flappy.


Project: http://git-wip-us.apache.org/repos/asf/phoenix/repo
Commit: http://git-wip-us.apache.org/repos/asf/phoenix/commit/71cc23c8
Tree: http://git-wip-us.apache.org/repos/asf/phoenix/tree/71cc23c8
Diff: http://git-wip-us.apache.org/repos/asf/phoenix/diff/71cc23c8

Branch: refs/heads/3.0
Commit: 71cc23c8fa03694db1816cf8e3f5d0bb3f391ccb
Parents: 19dc23a
Author: Jeffrey Zhong 
Authored: Fri Aug 15 14:02:51 2014 -0700
Committer: Jeffrey Zhong 
Committed: Fri Aug 15 14:02:51 2014 -0700

--
 .../apache/phoenix/coprocessor/MetaDataEndpointImpl.java| 2 ++
 .../apache/phoenix/coprocessor/MetaDataRegionObserver.java  | 9 ++---
 2 files changed, 8 insertions(+), 3 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/phoenix/blob/71cc23c8/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
--
diff --git 
a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
 
b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
index 07d4cc8..9cb3b89 100644
--- 
a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
+++ 
b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
@@ -1207,6 +1207,8 @@ public class MetaDataEndpointImpl extends 
BaseEndpointCoprocessor implements Met
 dataTableKey = SchemaUtil.getTableKey(tenantId, 
schemaName, dataTableKV.getValue());
 }
 if(dataTableKey != null) {
+// make a copy of tableMetadata
+tableMetadata = new ArrayList(tableMetadata);
 // insert an empty KV to trigger time stamp update on 
data table row
 Put p = new Put(dataTableKey);
 p.add(TABLE_FAMILY_BYTES, 
QueryConstants.EMPTY_COLUMN_BYTES, timeStamp, ByteUtil.EMPTY_BYTE_ARRAY);

http://git-wip-us.apache.org/repos/asf/phoenix/blob/71cc23c8/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataRegionObserver.java
--
diff --git 
a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataRegionObserver.java
 
b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataRegionObserver.java
index 2820e59..1526a98 100644
--- 
a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataRegionObserver.java
+++ 
b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataRegionObserver.java
@@ -24,6 +24,9 @@ import java.util.ArrayList;
 import java.util.List;
 import java.util.Timer;
 import java.util.TimerTask;
+import java.util.concurrent.ScheduledFuture;
+import java.util.concurrent.ScheduledThreadPoolExecutor;
+import java.util.concurrent.TimeUnit;
 import java.util.concurrent.atomic.AtomicInteger;
 
 import org.apache.commons.logging.Log;
@@ -64,14 +67,14 @@ import org.apache.phoenix.util.SchemaUtil;
  */
 public class MetaDataRegionObserver extends BaseRegionObserver {
 public static final Log LOG = 
LogFactory.getLog(MetaDataRegionObserver.class);
-protected Timer scheduleTimer = new Timer(true);
+protected ScheduledThreadPoolExecutor executor = new 
ScheduledThreadPoolExecutor(1);
 private boolean enableRebuildIndex = 
QueryServicesOptions.DEFAULT_INDEX_FAILURE_HANDLING_REBUILD;
 private long rebuildIndexTimeInterval = 
QueryServicesOptions.DEFAULT_INDEX_FAILURE_HANDLING_REBUILD_INTERVAL;
   
 @Override
 public void preClose(final ObserverContext c,
 boolean abortRequested) {
-scheduleTimer.cancel();
+executor.shutdownNow();
 
GlobalCache.getInstance(c.getEnvironment()).getMetaDataCache().invalidateAll();
 }
 
@@ -112,7 +115,7 @@ public class MetaDataRegionObserver extends 
BaseRegionObserver {
 // starts index rebuild schedule work
 BuildIndexScheduleTask task = new 
BuildIndexScheduleTask(e.getEnvironment());
 // run scheduled task every 10 secs
-scheduleTimer.schedule(task, 1, rebuildIndexTimeInterval);
+executor.scheduleAtFixedRate(task, 1, 
rebuildIndexTimeInterval, TimeUnit.MILLISECONDS);
 } catch (ClassNotFoundException ex) {
 LOG.error("BuildIndexScheduleTask cannot start!", ex);
 }



git commit: PHOENIX-1174 Rename and move properties using existing convention

2014-08-15 Thread jamestaylor
Repository: phoenix
Updated Branches:
  refs/heads/4.0 2a3a253c2 -> 5aa381516


PHOENIX-1174 Rename and move properties using existing convention


Project: http://git-wip-us.apache.org/repos/asf/phoenix/repo
Commit: http://git-wip-us.apache.org/repos/asf/phoenix/commit/5aa38151
Tree: http://git-wip-us.apache.org/repos/asf/phoenix/tree/5aa38151
Diff: http://git-wip-us.apache.org/repos/asf/phoenix/diff/5aa38151

Branch: refs/heads/4.0
Commit: 5aa38151681b0c9f7f15c7e520f1d1eea4821565
Parents: 2a3a253
Author: James Taylor 
Authored: Fri Aug 15 14:32:14 2014 -0700
Committer: James Taylor 
Committed: Fri Aug 15 14:32:14 2014 -0700

--
 .../phoenix/end2end/index/IndexHandlerIT.java   |  4 ++--
 .../ipc/PhoenixIndexRpcSchedulerFactory.java| 23 +---
 .../org/apache/phoenix/query/QueryServices.java |  9 
 .../phoenix/query/QueryServicesOptions.java | 11 --
 .../org/apache/phoenix/trace/util/Tracing.java  | 18 ++-
 .../PhoenixIndexRpcSchedulerFactoryTest.java|  5 +++--
 6 files changed, 35 insertions(+), 35 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/phoenix/blob/5aa38151/phoenix-core/src/it/java/org/apache/phoenix/end2end/index/IndexHandlerIT.java
--
diff --git 
a/phoenix-core/src/it/java/org/apache/phoenix/end2end/index/IndexHandlerIT.java 
b/phoenix-core/src/it/java/org/apache/phoenix/end2end/index/IndexHandlerIT.java
index 8536652..1507d6b 100644
--- 
a/phoenix-core/src/it/java/org/apache/phoenix/end2end/index/IndexHandlerIT.java
+++ 
b/phoenix-core/src/it/java/org/apache/phoenix/end2end/index/IndexHandlerIT.java
@@ -38,7 +38,7 @@ import org.apache.hadoop.hbase.ipc.RpcControllerFactory;
 import org.apache.hadoop.hbase.util.Bytes;
 import org.apache.phoenix.hbase.index.IndexQosRpcControllerFactory;
 import org.apache.phoenix.hbase.index.TableName;
-import org.apache.phoenix.hbase.index.ipc.PhoenixIndexRpcSchedulerFactory;
+import org.apache.phoenix.query.QueryServicesOptions;
 import org.junit.After;
 import org.junit.AfterClass;
 import org.junit.Before;
@@ -159,7 +159,7 @@ public class IndexHandlerIT {
 // check the counts on the rpc controller
 assertEquals("Didn't get the expected number of index priority 
writes!", 1,
 (int) CountingIndexClientRpcController.priorityCounts
-
.get(PhoenixIndexRpcSchedulerFactory.DEFAULT_INDEX_MIN_PRIORITY));
+.get(QueryServicesOptions.DEFAULT_INDEX_MIN_PRIORITY));
 
 table.close();
 }

http://git-wip-us.apache.org/repos/asf/phoenix/blob/5aa38151/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/ipc/PhoenixIndexRpcSchedulerFactory.java
--
diff --git 
a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/ipc/PhoenixIndexRpcSchedulerFactory.java
 
b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/ipc/PhoenixIndexRpcSchedulerFactory.java
index 500db7c..8e0b86f 100644
--- 
a/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/ipc/PhoenixIndexRpcSchedulerFactory.java
+++ 
b/phoenix-core/src/main/java/org/apache/phoenix/hbase/index/ipc/PhoenixIndexRpcSchedulerFactory.java
@@ -26,6 +26,8 @@ import org.apache.hadoop.hbase.ipc.RpcScheduler;
 import org.apache.hadoop.hbase.regionserver.RegionServerServices;
 import org.apache.hadoop.hbase.regionserver.RpcSchedulerFactory;
 import org.apache.hadoop.hbase.regionserver.SimpleRpcSchedulerFactory;
+import org.apache.phoenix.query.QueryServices;
+import org.apache.phoenix.query.QueryServicesOptions;
 
 import com.google.common.base.Preconditions;
 
@@ -37,21 +39,6 @@ public class PhoenixIndexRpcSchedulerFactory implements 
RpcSchedulerFactory {
 
 private static final Log LOG = 
LogFactory.getLog(PhoenixIndexRpcSchedulerFactory.class);
 
-private static final String INDEX_HANDLER_COUNT_KEY =
-"org.apache.phoenix.regionserver.index.handler.count";
-private static final int DEFAULT_INDEX_HANDLER_COUNT = 30;
-
-/**
- * HConstants#HIGH_QOS is the max we will see to a standard table. We go 
higher to differentiate
- * and give some room for things in the middle
- */
-public static final int DEFAULT_INDEX_MIN_PRIORITY = 1000;
-public static final int DEFAULT_INDEX_MAX_PRIORITY = 1050;
-public static final String MIN_INDEX_PRIOIRTY_KEY =
-"org.apache.phoenix.regionserver.index.priority.min";
-public static final String MAX_INDEX_PRIOIRTY_KEY =
-"org.apache.phoenix.regionserver.index.priority.max";
-
 private static final String VERSION_TOO_OLD_FOR_INDEX_RPC =
 "Running an older version of HBase (less than 0.98.4), Phoenix 
index RPC handling cannot be enabled.";
 
@@ -75,9 +62,9 @@ public class PhoenixIndexRpcSche

svn commit: r1618279 - in /phoenix/site: publish/ publish/language/ source/src/site/

2014-08-15 Thread mujtaba
Author: mujtaba
Date: Fri Aug 15 21:35:37 2014
New Revision: 1618279

URL: http://svn.apache.org/r1618279
Log:
Fix Phoenix site search URL

Modified:
phoenix/site/publish/Phoenix-in-15-minutes-or-less.html
phoenix/site/publish/array_type.html
phoenix/site/publish/building.html
phoenix/site/publish/building_website.html
phoenix/site/publish/bulk_dataload.html
phoenix/site/publish/contributing.html
phoenix/site/publish/download.html
phoenix/site/publish/dynamic_columns.html
phoenix/site/publish/faq.html
phoenix/site/publish/flume.html
phoenix/site/publish/index.html
phoenix/site/publish/issues.html
phoenix/site/publish/joins.html
phoenix/site/publish/language/datatypes.html
phoenix/site/publish/language/functions.html
phoenix/site/publish/language/index.html
phoenix/site/publish/mailing_list.html
phoenix/site/publish/multi-tenancy.html
phoenix/site/publish/paged.html
phoenix/site/publish/performance.html
phoenix/site/publish/phoenix_on_emr.html
phoenix/site/publish/pig_integration.html
phoenix/site/publish/recent.html
phoenix/site/publish/resources.html
phoenix/site/publish/roadmap.html
phoenix/site/publish/salted.html
phoenix/site/publish/secondary_indexing.html
phoenix/site/publish/sequences.html
phoenix/site/publish/skip_scan.html
phoenix/site/publish/source.html
phoenix/site/publish/team.html
phoenix/site/publish/tracing.html
phoenix/site/publish/tuning.html
phoenix/site/publish/upgrade_from_2_2.html
phoenix/site/publish/views.html
phoenix/site/source/src/site/site.xml

Modified: phoenix/site/publish/Phoenix-in-15-minutes-or-less.html
URL: 
http://svn.apache.org/viewvc/phoenix/site/publish/Phoenix-in-15-minutes-or-less.html?rev=1618279&r1=1618278&r2=1618279&view=diff
==
--- phoenix/site/publish/Phoenix-in-15-minutes-or-less.html (original)
+++ phoenix/site/publish/Phoenix-in-15-minutes-or-less.html Fri Aug 15 21:35:37 
2014
@@ -1,7 +1,7 @@
 
 
 
 
@@ -330,7 +330,7 @@



-   https://www.google.com/search"; method="get">
+   https://www.google.com/search"; method="get">




Modified: phoenix/site/publish/array_type.html
URL: 
http://svn.apache.org/viewvc/phoenix/site/publish/array_type.html?rev=1618279&r1=1618278&r2=1618279&view=diff
==
--- phoenix/site/publish/array_type.html (original)
+++ phoenix/site/publish/array_type.html Fri Aug 15 21:35:37 2014
@@ -1,7 +1,7 @@
 
 
 
 
@@ -356,7 +356,7 @@ SELECT region_name FROM regions WHERE '9



-   https://www.google.com/search"; method="get">
+   https://www.google.com/search"; method="get">




Modified: phoenix/site/publish/building.html
URL: 
http://svn.apache.org/viewvc/phoenix/site/publish/building.html?rev=1618279&r1=1618278&r2=1618279&view=diff
==
--- phoenix/site/publish/building.html (original)
+++ phoenix/site/publish/building.html Fri Aug 15 21:35:37 2014
@@ -1,7 +1,7 @@
 
 
 
 
@@ -335,7 +335,7 @@



-   https://www.google.com/search"; method="get">
+   https://www.google.com/search"; method="get">




Modified: phoenix/site/publish/building_website.html
URL: 
http://svn.apache.org/viewvc/phoenix/site/publish/building_website.html?rev=1618279&r1=1618278&r2=1618279&view=diff
==
--- phoenix/site/publish/building_website.html (original)
+++ phoenix/site/publish/building_website.html Fri Aug 15 21:35:37 2014
@@ -1,7 +1,7 @@
 
 
 
 
@@ -287,7 +287,7 @@



-   https://www.google.com/search"; method="get">
+   https://www.google.com/search"; method="get">




Modified: phoenix/site/publish/bulk_dataload.html
URL: 
http://svn.apache.org/viewvc/phoenix/site/publish/bulk_dataload.html?rev=1618279&r1=1618278&r2=1618279&view=diff
===

Jenkins build is back to normal : Phoenix | Master | Hadoop1 #332

2014-08-15 Thread Apache Jenkins Server
See 



Jenkins build is back to normal : Phoenix | 4.0 | Hadoop1 #278

2014-08-15 Thread Apache Jenkins Server
See 



Jenkins build is back to normal : Phoenix | 4.0 | Hadoop2 #36

2014-08-15 Thread Apache Jenkins Server
See 



git commit: Rename com.salesforce tracing config prop to org.apache

2014-08-15 Thread jamestaylor
Repository: phoenix
Updated Branches:
  refs/heads/master acd35f0eb -> dfcde1046


Rename com.salesforce tracing config prop to org.apache


Project: http://git-wip-us.apache.org/repos/asf/phoenix/repo
Commit: http://git-wip-us.apache.org/repos/asf/phoenix/commit/dfcde104
Tree: http://git-wip-us.apache.org/repos/asf/phoenix/tree/dfcde104
Diff: http://git-wip-us.apache.org/repos/asf/phoenix/diff/dfcde104

Branch: refs/heads/master
Commit: dfcde10464a028c8acf44b9ce662f0fd19ca265d
Parents: acd35f0
Author: James Taylor 
Authored: Fri Aug 15 09:59:50 2014 -0700
Committer: James Taylor 
Committed: Fri Aug 15 09:59:50 2014 -0700

--
 .../src/main/java/org/apache/phoenix/trace/util/Tracing.java   | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/phoenix/blob/dfcde104/phoenix-core/src/main/java/org/apache/phoenix/trace/util/Tracing.java
--
diff --git 
a/phoenix-core/src/main/java/org/apache/phoenix/trace/util/Tracing.java 
b/phoenix-core/src/main/java/org/apache/phoenix/trace/util/Tracing.java
index 5913cfb..3367bce 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/trace/util/Tracing.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/trace/util/Tracing.java
@@ -59,7 +59,7 @@ public class Tracing {
 // Constants for for configuring tracing
 public static final String TRACING_LEVEL_KEY = 
"org.apache.phoenix.trace.frequency";
 protected static final String PROBABILITY_THRESHOLD_KEY =
-"com.salesforce.phoenix.trace.probability.threshold";
+"org.apache.phoenix.trace.probability.threshold";
 
 /**
  * We always trace on the server, assuming the client has requested 
tracing on the request



git commit: Rename com.salesforce tracing config prop to org.apache

2014-08-15 Thread jamestaylor
Repository: phoenix
Updated Branches:
  refs/heads/4.0 fcdcd697c -> 2a3a253c2


Rename com.salesforce tracing config prop to org.apache


Project: http://git-wip-us.apache.org/repos/asf/phoenix/repo
Commit: http://git-wip-us.apache.org/repos/asf/phoenix/commit/2a3a253c
Tree: http://git-wip-us.apache.org/repos/asf/phoenix/tree/2a3a253c
Diff: http://git-wip-us.apache.org/repos/asf/phoenix/diff/2a3a253c

Branch: refs/heads/4.0
Commit: 2a3a253c2ede05d6a7c0cab03db93e9fe13efa73
Parents: fcdcd69
Author: James Taylor 
Authored: Fri Aug 15 09:58:15 2014 -0700
Committer: James Taylor 
Committed: Fri Aug 15 09:58:15 2014 -0700

--
 .../src/main/java/org/apache/phoenix/trace/util/Tracing.java   | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/phoenix/blob/2a3a253c/phoenix-core/src/main/java/org/apache/phoenix/trace/util/Tracing.java
--
diff --git 
a/phoenix-core/src/main/java/org/apache/phoenix/trace/util/Tracing.java 
b/phoenix-core/src/main/java/org/apache/phoenix/trace/util/Tracing.java
index 5913cfb..3367bce 100644
--- a/phoenix-core/src/main/java/org/apache/phoenix/trace/util/Tracing.java
+++ b/phoenix-core/src/main/java/org/apache/phoenix/trace/util/Tracing.java
@@ -59,7 +59,7 @@ public class Tracing {
 // Constants for for configuring tracing
 public static final String TRACING_LEVEL_KEY = 
"org.apache.phoenix.trace.frequency";
 protected static final String PROBABILITY_THRESHOLD_KEY =
-"com.salesforce.phoenix.trace.probability.threshold";
+"org.apache.phoenix.trace.probability.threshold";
 
 /**
  * We always trace on the server, assuming the client has requested 
tracing on the request



Build failed in Jenkins: Phoenix | 4.0 | Hadoop1 #277

2014-08-15 Thread Apache Jenkins Server
See 

Changes:

[jtaylor] PHOENIX-1047 Auto cast - add/sub decimal constant and integer

--
Started by an SCM change
Started by an SCM change
Building remotely on ubuntu3 (Ubuntu ubuntu) in workspace 

 > git rev-parse --is-inside-work-tree
Fetching changes from the remote Git repository
 > git config remote.origin.url 
 > https://git-wip-us.apache.org/repos/asf/phoenix.git
Fetching upstream changes from 
https://git-wip-us.apache.org/repos/asf/phoenix.git
 > git --version
 > git fetch --tags --progress 
 > https://git-wip-us.apache.org/repos/asf/phoenix.git 
 > +refs/heads/*:refs/remotes/origin/*
 > git rev-parse origin/4.0^{commit}
Checking out Revision fcdcd697ccf13cd377980a186c32fb8b4121d1a1 (origin/4.0)
 > git config core.sparsecheckout
 > git checkout -f fcdcd697ccf13cd377980a186c32fb8b4121d1a1
 > git rev-list b07658e8791cebf59ec45beeac56ec7ca5252a4d
No emails were triggered.
FATAL: Unable to produce a script file
java.io.IOException: Failed to create a temp file on 

at hudson.FilePath.createTextTempFile(FilePath.java:1265)
at 
hudson.tasks.CommandInterpreter.createScriptFile(CommandInterpreter.java:144)
at hudson.tasks.CommandInterpreter.perform(CommandInterpreter.java:82)
at hudson.tasks.CommandInterpreter.perform(CommandInterpreter.java:66)
at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
at 
hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:804)
at hudson.model.Build$BuildExecution.build(Build.java:199)
at hudson.model.Build$BuildExecution.doRun(Build.java:160)
at 
hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:585)
at hudson.model.Run.execute(Run.java:1676)
at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
at hudson.model.ResourceController.execute(ResourceController.java:88)
at hudson.model.Executor.run(Executor.java:231)
Caused by: java.io.IOException: remote file operation failed: 
 at 
hudson.remoting.Channel@29e76aff:ubuntu3
at hudson.FilePath.act(FilePath.java:910)
at hudson.FilePath.act(FilePath.java:887)
at hudson.FilePath.createTextTempFile(FilePath.java:1239)
... 12 more
Caused by: java.io.IOException: No space left on device
at java.io.FileOutputStream.writeBytes(Native Method)
at java.io.FileOutputStream.write(FileOutputStream.java:345)
at sun.nio.cs.StreamEncoder.writeBytes(StreamEncoder.java:221)
at sun.nio.cs.StreamEncoder.implClose(StreamEncoder.java:316)
at sun.nio.cs.StreamEncoder.close(StreamEncoder.java:149)
at java.io.OutputStreamWriter.close(OutputStreamWriter.java:233)
at hudson.FilePath$15.invoke(FilePath.java:1258)
at hudson.FilePath$15.invoke(FilePath.java:1239)
at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2462)
at hudson.remoting.UserRequest.perform(UserRequest.java:118)
at hudson.remoting.UserRequest.perform(UserRequest.java:48)
at hudson.remoting.Request$2.run(Request.java:328)
at 
hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Build step 'Execute shell' marked build as failure
Archiving artifacts
Sending artifact delta relative to Phoenix | 4.0 | Hadoop1 #276
Archived 566 artifacts
Archive block size is 32768
Received 1118 blocks and 198799185 bytes
Compression is 15.6%
Took 59 sec
Recording test results


Build failed in Jenkins: Phoenix | Master | Hadoop1 #331

2014-08-15 Thread Apache Jenkins Server
See 

Changes:

[jtaylor] PHOENIX-1047 Auto cast - add/sub decimal constant and integer

--
Started by an SCM change
Started by an SCM change
Building remotely on ubuntu3 (Ubuntu ubuntu) in workspace 

 > git rev-parse --is-inside-work-tree
Fetching changes from the remote Git repository
 > git config remote.origin.url 
 > https://git-wip-us.apache.org/repos/asf/phoenix.git
Fetching upstream changes from 
https://git-wip-us.apache.org/repos/asf/phoenix.git
 > git --version
 > git fetch --tags --progress 
 > https://git-wip-us.apache.org/repos/asf/phoenix.git 
 > +refs/heads/*:refs/remotes/origin/*
 > git rev-parse origin/master^{commit}
Checking out Revision acd35f0ebbb3cf3151741832dc3f3e08a585318c (origin/master)
 > git config core.sparsecheckout
 > git checkout -f acd35f0ebbb3cf3151741832dc3f3e08a585318c
 > git rev-list c85d4c6ad145babcd5eca2fde1dc632071105b77
No emails were triggered.
FATAL: Unable to produce a script file
java.io.IOException: Failed to create a temp file on 

at hudson.FilePath.createTextTempFile(FilePath.java:1265)
at 
hudson.tasks.CommandInterpreter.createScriptFile(CommandInterpreter.java:144)
at hudson.tasks.CommandInterpreter.perform(CommandInterpreter.java:82)
at hudson.tasks.CommandInterpreter.perform(CommandInterpreter.java:66)
at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
at 
hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:804)
at hudson.model.Build$BuildExecution.build(Build.java:199)
at hudson.model.Build$BuildExecution.doRun(Build.java:160)
at 
hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:585)
at hudson.model.Run.execute(Run.java:1676)
at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
at hudson.model.ResourceController.execute(ResourceController.java:88)
at hudson.model.Executor.run(Executor.java:231)
Caused by: java.io.IOException: remote file operation failed: 
 at 
hudson.remoting.Channel@29e76aff:ubuntu3
at hudson.FilePath.act(FilePath.java:910)
at hudson.FilePath.act(FilePath.java:887)
at hudson.FilePath.createTextTempFile(FilePath.java:1239)
... 12 more
Caused by: java.io.IOException: No space left on device
at java.io.FileOutputStream.writeBytes(Native Method)
at java.io.FileOutputStream.write(FileOutputStream.java:345)
at sun.nio.cs.StreamEncoder.writeBytes(StreamEncoder.java:221)
at sun.nio.cs.StreamEncoder.implClose(StreamEncoder.java:316)
at sun.nio.cs.StreamEncoder.close(StreamEncoder.java:149)
at java.io.OutputStreamWriter.close(OutputStreamWriter.java:233)
at hudson.FilePath$15.invoke(FilePath.java:1258)
at hudson.FilePath$15.invoke(FilePath.java:1239)
at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2462)
at hudson.remoting.UserRequest.perform(UserRequest.java:118)
at hudson.remoting.UserRequest.perform(UserRequest.java:48)
at hudson.remoting.Request$2.run(Request.java:328)
at 
hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Build step 'Execute shell' marked build as failure
Archiving artifacts
Sending artifact delta relative to Phoenix | Master | Hadoop1 #330
Archived 573 artifacts
Archive block size is 32768
Received 1069 blocks and 180826303 bytes
Compression is 16.2%
Took 56 sec
Updating PHOENIX-1047
Recording test results


Build failed in Jenkins: Phoenix | 3.0 | Hadoop1 #186

2014-08-15 Thread Apache Jenkins Server
See 

Changes:

[jtaylor] PHOENIX-1047 Auto cast - add/sub decimal constant and integer

--
[...truncated 149 lines...]
---

---
 T E S T S
---
Java HotSpot(TM) Server VM warning: Insufficient space for shared memory file:
   /tmp/hsperfdata_jenkins/25218
Try using the -Djava.io.tmpdir= option to select an alternate temp location.
Java HotSpot(TM) Server VM warning: Insufficient space for shared memory file:
   /tmp/hsperfdata_jenkins/25215
Try using the -Djava.io.tmpdir= option to select an alternate temp location.
Java HotSpot(TM) Server VM warning: Insufficient space for shared memory file:
   /tmp/hsperfdata_jenkins/25219
Try using the -Djava.io.tmpdir= option to select an alternate temp location.
Running org.apache.phoenix.client.TestClientKeyValueLocal
Running org.apache.phoenix.arithmetic.ArithmeticOperationTest
Running org.apache.phoenix.parse.QueryParserTest
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.253 sec - in 
org.apache.phoenix.client.TestClientKeyValueLocal
Running org.apache.phoenix.query.ConnectionlessTest
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.381 sec - in 
org.apache.phoenix.arithmetic.ArithmeticOperationTest
Running org.apache.phoenix.query.HBaseFactoryProviderTest
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.005 sec - in 
org.apache.phoenix.query.HBaseFactoryProviderTest
Running org.apache.phoenix.index.IndexMaintainerTest
Tests run: 46, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.563 sec - in 
org.apache.phoenix.parse.QueryParserTest
Running org.apache.phoenix.filter.SkipScanFilterIntersectTest
Tests run: 17, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.05 sec - in 
org.apache.phoenix.filter.SkipScanFilterIntersectTest
Running org.apache.phoenix.filter.SkipScanFilterTest
Tests run: 13, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.038 sec - in 
org.apache.phoenix.filter.SkipScanFilterTest
Running org.apache.phoenix.compile.WhereOptimizerTest
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.731 sec - in 
org.apache.phoenix.query.ConnectionlessTest
Running org.apache.phoenix.compile.ViewCompilerTest
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.364 sec - in 
org.apache.phoenix.compile.ViewCompilerTest
Running org.apache.phoenix.compile.QueryOptimizerTest
Tests run: 94, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.781 sec - in 
org.apache.phoenix.compile.WhereOptimizerTest
Running org.apache.phoenix.compile.WhereCompilerTest
Tests run: 41, Failures: 0, Errors: 0, Skipped: 1, Time elapsed: 0.234 sec - in 
org.apache.phoenix.compile.WhereCompilerTest
Running org.apache.phoenix.compile.ScanRangesTest
Tests run: 34, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.027 sec - in 
org.apache.phoenix.compile.ScanRangesTest
Running org.apache.phoenix.compile.JoinQueryCompilerTest
Tests run: 25, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.411 sec - in 
org.apache.phoenix.compile.QueryOptimizerTest
Running org.apache.phoenix.compile.HavingCompilerTest
Tests run: 23, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.382 sec - in 
org.apache.phoenix.index.IndexMaintainerTest
Running org.apache.phoenix.compile.QueryCompilerTest
Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.105 sec - in 
org.apache.phoenix.compile.HavingCompilerTest
Running org.apache.phoenix.compile.StatementHintsCompilationTest
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.184 sec - in 
org.apache.phoenix.compile.JoinQueryCompilerTest
Running org.apache.phoenix.compile.TenantSpecificViewIndexCompileTest
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.109 sec - in 
org.apache.phoenix.compile.StatementHintsCompilationTest
Running org.apache.phoenix.compile.LimitCompilerTest
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.083 sec - in 
org.apache.phoenix.compile.LimitCompilerTest
Running org.apache.phoenix.compile.QueryMetaDataTest
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.225 sec - in 
org.apache.phoenix.compile.TenantSpecificViewIndexCompileTest
Running org.apache.phoenix.compile.SelectStatementRewriterTest
Tests run: 83, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.427 sec - in 
org.apache.phoenix.compile.QueryCompilerTest
Running org.apache.phoenix.mapreduce.CsvBulkImportUtilTest
Tests run: 32, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.129 sec - in 
org.apache.phoenix.compile.QueryMetaDataTest
Running org.apache.phoenix.mapreduce.CsvToKeyValueMapperTest
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.091 sec - in 
org.apache.phoenix.compile.SelectStatementRewriterTest
Running org.apache.phoenix.mapr

Build failed in Jenkins: Phoenix | 4.0 | Hadoop2 #34

2014-08-15 Thread Apache Jenkins Server
See 

Changes:

[jtaylor] PHOENIX-1171 Dropping the index is not verifying the associated table

[jtaylor] PHOENIX-1047 Auto cast - add/sub decimal constant and integer

--
Started by an SCM change
Started by an SCM change
Started by an SCM change
Building remotely on ubuntu3 (Ubuntu ubuntu) in workspace 

 > git rev-parse --is-inside-work-tree
Fetching changes from the remote Git repository
 > git config remote.origin.url git://git.apache.org/phoenix.git/
Fetching upstream changes from git://git.apache.org/phoenix.git/
 > git --version
 > git fetch --tags --progress git://git.apache.org/phoenix.git/ 
 > +refs/heads/*:refs/remotes/origin/*
 > git rev-parse origin/4.0^{commit}
Checking out Revision fcdcd697ccf13cd377980a186c32fb8b4121d1a1 (origin/4.0)
 > git config core.sparsecheckout
 > git checkout -f fcdcd697ccf13cd377980a186c32fb8b4121d1a1
 > git rev-list 2bcc1d14889685ecc6187fb33684aece74de51c3
No emails were triggered.
FATAL: Unable to produce a script file
java.io.IOException: Failed to create a temp file on 

at hudson.FilePath.createTextTempFile(FilePath.java:1265)
at 
hudson.tasks.CommandInterpreter.createScriptFile(CommandInterpreter.java:144)
at hudson.tasks.CommandInterpreter.perform(CommandInterpreter.java:82)
at hudson.tasks.CommandInterpreter.perform(CommandInterpreter.java:66)
at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
at 
hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:804)
at hudson.model.Build$BuildExecution.build(Build.java:199)
at hudson.model.Build$BuildExecution.doRun(Build.java:160)
at 
hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:585)
at hudson.model.Run.execute(Run.java:1676)
at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
at hudson.model.ResourceController.execute(ResourceController.java:88)
at hudson.model.Executor.run(Executor.java:231)
Caused by: java.io.IOException: remote file operation failed: 
 at 
hudson.remoting.Channel@29e76aff:ubuntu3
at hudson.FilePath.act(FilePath.java:910)
at hudson.FilePath.act(FilePath.java:887)
at hudson.FilePath.createTextTempFile(FilePath.java:1239)
... 12 more
Caused by: java.io.IOException: No space left on device
at java.io.FileOutputStream.writeBytes(Native Method)
at java.io.FileOutputStream.write(FileOutputStream.java:345)
at sun.nio.cs.StreamEncoder.writeBytes(StreamEncoder.java:221)
at sun.nio.cs.StreamEncoder.implClose(StreamEncoder.java:316)
at sun.nio.cs.StreamEncoder.close(StreamEncoder.java:149)
at java.io.OutputStreamWriter.close(OutputStreamWriter.java:233)
at hudson.FilePath$15.invoke(FilePath.java:1258)
at hudson.FilePath$15.invoke(FilePath.java:1239)
at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2462)
at hudson.remoting.UserRequest.perform(UserRequest.java:118)
at hudson.remoting.UserRequest.perform(UserRequest.java:48)
at hudson.remoting.Request$2.run(Request.java:328)
at 
hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Build step 'Execute shell' marked build as failure
Archiving artifacts
Sending artifact delta relative to Phoenix | 4.0 | Hadoop2 #32
ERROR: Failed to archive 
{phoenix-core/target/failsafe-reports/TEST-org.apache.phoenix.end2end.ModulusExpressionIT.xml=phoenix-core/target/failsafe-reports/TEST-org.apache.phoenix.end2end.ModulusExpressionIT.xml,
 
phoenix-core/target/surefire-reports/org.apache.phoenix.util.csv.CsvUpsertExecutorTest-output.txt=phoenix-core/target/surefire-reports/org.apache.phoenix.util.csv.CsvUpsertExecutorTest-output.txt,
 
phoenix-core/target/surefire-reports/org.apache.phoenix.compile.QueryCompilerTest-output.txt=phoenix-core/target/surefire-reports/org.apache.phoenix.compile.QueryCompilerTest-output.txt,
 
phoenix-core/target/failsafe-reports/org.apache.phoenix.hbase.index.covered.example.EndtoEndIndexingWithCompressionIT.txt=phoenix-core/target/failsafe-reports/org.apache.phoenix.hbase.index.covered.example.EndtoEndIndexingWithCompressionIT.txt,
 
phoenix-core/target/surefire-reports/TEST-org.apache.phoenix.expression.LikeExpressionTest.xml=phoenix-core/target/surefire-reports/TEST-org.apache.phoenix.expression.LikeExpressionTest.xml,
 
phoenix-assembl