git commit: PHOENIX-1208 Check for existence of views doesn't take into account the fact that SYSTEM.CATALOG could be split across regions

2014-08-28 Thread jamestaylor
Repository: phoenix
Updated Branches:
  refs/heads/3.0 e1ba4dec4 -> 87b223fcc


PHOENIX-1208 Check for existence of views doesn't take into account the fact 
that SYSTEM.CATALOG could be split across regions

Conflicts:

phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java


Project: http://git-wip-us.apache.org/repos/asf/phoenix/repo
Commit: http://git-wip-us.apache.org/repos/asf/phoenix/commit/87b223fc
Tree: http://git-wip-us.apache.org/repos/asf/phoenix/tree/87b223fc
Diff: http://git-wip-us.apache.org/repos/asf/phoenix/diff/87b223fc

Branch: refs/heads/3.0
Commit: 87b223fcc88e030f4e85677f49fc6004ca13c78e
Parents: e1ba4de
Author: James Taylor 
Authored: Thu Aug 28 22:29:50 2014 -0700
Committer: James Taylor 
Committed: Thu Aug 28 22:44:20 2014 -0700

--
 .../coprocessor/MetaDataEndpointImpl.java   | 67 +---
 .../phoenix/coprocessor/SuffixFilter.java   | 45 +
 2 files changed, 61 insertions(+), 51 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/phoenix/blob/87b223fc/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
--
diff --git 
a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
 
b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
index 9cb3b89..b7343d1 100644
--- 
a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
+++ 
b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
@@ -17,8 +17,6 @@
  */
 package org.apache.phoenix.coprocessor;
 
-import static com.google.common.collect.Lists.newArrayList;
-import static org.apache.hadoop.hbase.filter.CompareFilter.CompareOp.EQUAL;
 import static org.apache.phoenix.jdbc.PhoenixDatabaseMetaData.ARRAY_SIZE_BYTES;
 import static 
org.apache.phoenix.jdbc.PhoenixDatabaseMetaData.COLUMN_COUNT_BYTES;
 import static 
org.apache.phoenix.jdbc.PhoenixDatabaseMetaData.COLUMN_NAME_INDEX;
@@ -54,8 +52,6 @@ import static org.apache.phoenix.schema.PTableType.INDEX;
 import static org.apache.phoenix.util.SchemaUtil.getVarCharLength;
 import static org.apache.phoenix.util.SchemaUtil.getVarChars;
 
-import java.io.DataInput;
-import java.io.DataOutput;
 import java.io.IOException;
 import java.sql.ResultSetMetaData;
 import java.sql.SQLException;
@@ -68,14 +64,16 @@ import org.apache.hadoop.hbase.HConstants;
 import org.apache.hadoop.hbase.KeyValue;
 import org.apache.hadoop.hbase.client.Delete;
 import org.apache.hadoop.hbase.client.Get;
+import org.apache.hadoop.hbase.client.HTableInterface;
 import org.apache.hadoop.hbase.client.Mutation;
 import org.apache.hadoop.hbase.client.Put;
 import org.apache.hadoop.hbase.client.Result;
+import org.apache.hadoop.hbase.client.ResultScanner;
 import org.apache.hadoop.hbase.client.Scan;
 import org.apache.hadoop.hbase.coprocessor.BaseEndpointCoprocessor;
 import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment;
+import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
 import org.apache.hadoop.hbase.filter.Filter;
-import org.apache.hadoop.hbase.filter.FilterBase;
 import org.apache.hadoop.hbase.filter.FilterList;
 import org.apache.hadoop.hbase.filter.FirstKeyOnlyFilter;
 import org.apache.hadoop.hbase.filter.SingleColumnValueFilter;
@@ -596,7 +594,7 @@ public class MetaDataEndpointImpl extends 
BaseEndpointCoprocessor implements Met
 }
 }
 
-private static final byte[] PHYSICAL_TABLE_BYTES = new byte[] 
{PTable.LinkType.PHYSICAL_TABLE.getSerializedValue()};
+protected static final byte[] PHYSICAL_TABLE_BYTES = new byte[] 
{PTable.LinkType.PHYSICAL_TABLE.getSerializedValue()};
 /**
  * @param tableName parent table's name
  * @return true if there exist a table that use this table as their base 
table.
@@ -616,21 +614,25 @@ public class MetaDataEndpointImpl extends 
BaseEndpointCoprocessor implements Met
 scan.setStartRow(startRow);
 scan.setStopRow(stopRow);
 }
-SingleColumnValueFilter linkFilter = new 
SingleColumnValueFilter(TABLE_FAMILY_BYTES, LINK_TYPE_BYTES, EQUAL, 
PHYSICAL_TABLE_BYTES);
+SingleColumnValueFilter linkFilter = new 
SingleColumnValueFilter(TABLE_FAMILY_BYTES, LINK_TYPE_BYTES, CompareOp.EQUAL, 
PHYSICAL_TABLE_BYTES);
 linkFilter.setFilterIfMissing(true);
 byte[] suffix = ByteUtil.concat(QueryConstants.SEPARATOR_BYTE_ARRAY, 
SchemaUtil.getTableNameAsBytes(schemaName, tableName));
 SuffixFilter rowFilter = new SuffixFilter(suffix);
 Filter filter = new FilterList(linkFilter, rowFilter);
 scan.setFilter(filter);
 scan.addColumn(TABLE_FAMILY_BYTES, LINK_TYPE_BYTES);
-RegionScanner scanner = region.getScanner(scan);

git commit: PHOENIX-1208 Check for existence of views doesn't take into account the fact that SYSTEM.CATALOG could be split across regions

2014-08-28 Thread jamestaylor
Repository: phoenix
Updated Branches:
  refs/heads/master bd174c1c9 -> 867af78dd


PHOENIX-1208 Check for existence of views doesn't take into account the fact 
that SYSTEM.CATALOG could be split across regions


Project: http://git-wip-us.apache.org/repos/asf/phoenix/repo
Commit: http://git-wip-us.apache.org/repos/asf/phoenix/commit/867af78d
Tree: http://git-wip-us.apache.org/repos/asf/phoenix/tree/867af78d
Diff: http://git-wip-us.apache.org/repos/asf/phoenix/diff/867af78d

Branch: refs/heads/master
Commit: 867af78dda90910f07dee0ac9c95cd4a7f50a749
Parents: bd174c1
Author: James Taylor 
Authored: Thu Aug 28 22:29:50 2014 -0700
Committer: James Taylor 
Committed: Thu Aug 28 22:31:06 2014 -0700

--
 .../coprocessor/MetaDataEndpointImpl.java   | 73 
 .../phoenix/coprocessor/SuffixFilter.java   | 50 ++
 2 files changed, 79 insertions(+), 44 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/phoenix/blob/867af78d/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
--
diff --git 
a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
 
b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
index 2f7b34f..020a3c9 100644
--- 
a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
+++ 
b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
@@ -17,8 +17,6 @@
  */
 package org.apache.phoenix.coprocessor;
 
-import static com.google.common.collect.Lists.newArrayList;
-import static org.apache.hadoop.hbase.filter.CompareFilter.CompareOp.EQUAL;
 import static org.apache.phoenix.jdbc.PhoenixDatabaseMetaData.ARRAY_SIZE_BYTES;
 import static 
org.apache.phoenix.jdbc.PhoenixDatabaseMetaData.COLUMN_COUNT_BYTES;
 import static 
org.apache.phoenix.jdbc.PhoenixDatabaseMetaData.COLUMN_NAME_INDEX;
@@ -73,15 +71,18 @@ import org.apache.hadoop.hbase.KeyValue;
 import org.apache.hadoop.hbase.KeyValue.Type;
 import org.apache.hadoop.hbase.client.Delete;
 import org.apache.hadoop.hbase.client.Get;
+import org.apache.hadoop.hbase.client.HTableInterface;
+import org.apache.hadoop.hbase.client.HTablePool;
 import org.apache.hadoop.hbase.client.Mutation;
 import org.apache.hadoop.hbase.client.Put;
 import org.apache.hadoop.hbase.client.Result;
+import org.apache.hadoop.hbase.client.ResultScanner;
 import org.apache.hadoop.hbase.client.Scan;
 import org.apache.hadoop.hbase.coprocessor.CoprocessorException;
 import org.apache.hadoop.hbase.coprocessor.CoprocessorService;
 import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment;
+import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
 import org.apache.hadoop.hbase.filter.Filter;
-import org.apache.hadoop.hbase.filter.FilterBase;
 import org.apache.hadoop.hbase.filter.FilterList;
 import org.apache.hadoop.hbase.filter.FirstKeyOnlyFilter;
 import org.apache.hadoop.hbase.filter.SingleColumnValueFilter;
@@ -857,12 +858,13 @@ public class MetaDataEndpointImpl extends 
MetaDataProtocol implements Coprocesso
 locks.add(rowLock);
 }
 
-private static final byte[] PHYSICAL_TABLE_BYTES = new byte[] 
{PTable.LinkType.PHYSICAL_TABLE.getSerializedValue()};
+protected static final byte[] PHYSICAL_TABLE_BYTES = new byte[] 
{PTable.LinkType.PHYSICAL_TABLE.getSerializedValue()};
 /**
  * @param tableName parent table's name
  * @return true if there exist a table that use this table as their base 
table.
  * TODO: should we pass a timestamp here?
  */
+@SuppressWarnings("deprecation")
 private boolean hasViews(HRegion region, byte[] tenantId, PTable table) 
throws IOException {
 byte[] schemaName = table.getSchemaName().getBytes();
 byte[] tableName = table.getTableName().getBytes();
@@ -877,21 +879,36 @@ public class MetaDataEndpointImpl extends 
MetaDataProtocol implements Coprocesso
 scan.setStartRow(startRow);
 scan.setStopRow(stopRow);
 }
-SingleColumnValueFilter linkFilter = new 
SingleColumnValueFilter(TABLE_FAMILY_BYTES, LINK_TYPE_BYTES, EQUAL, 
PHYSICAL_TABLE_BYTES);
+SingleColumnValueFilter linkFilter = new 
SingleColumnValueFilter(TABLE_FAMILY_BYTES, LINK_TYPE_BYTES, CompareOp.EQUAL, 
PHYSICAL_TABLE_BYTES);
 linkFilter.setFilterIfMissing(true);
 byte[] suffix = ByteUtil.concat(QueryConstants.SEPARATOR_BYTE_ARRAY, 
SchemaUtil.getTableNameAsBytes(schemaName, tableName));
 SuffixFilter rowFilter = new SuffixFilter(suffix);
 Filter filter = new FilterList(linkFilter, rowFilter);
 scan.setFilter(filter);
 scan.addColumn(TABLE_FAMILY_BYTES, LINK_TYPE_BYTES);
-RegionScanner scanner = region.getScanner(scan);
+

git commit: PHOENIX-1208 Check for existence of views doesn't take into account the fact that SYSTEM.CATALOG could be split across regions

2014-08-28 Thread jamestaylor
Repository: phoenix
Updated Branches:
  refs/heads/4.0 13903e055 -> 6fb2b22b9


PHOENIX-1208 Check for existence of views doesn't take into account the fact 
that SYSTEM.CATALOG could be split across regions


Project: http://git-wip-us.apache.org/repos/asf/phoenix/repo
Commit: http://git-wip-us.apache.org/repos/asf/phoenix/commit/6fb2b22b
Tree: http://git-wip-us.apache.org/repos/asf/phoenix/tree/6fb2b22b
Diff: http://git-wip-us.apache.org/repos/asf/phoenix/diff/6fb2b22b

Branch: refs/heads/4.0
Commit: 6fb2b22b9a30ecf74b608cc1d6081b7889763f20
Parents: 13903e0
Author: James Taylor 
Authored: Thu Aug 28 22:29:50 2014 -0700
Committer: James Taylor 
Committed: Thu Aug 28 22:29:50 2014 -0700

--
 .../coprocessor/MetaDataEndpointImpl.java   | 73 
 .../phoenix/coprocessor/SuffixFilter.java   | 50 ++
 2 files changed, 79 insertions(+), 44 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/phoenix/blob/6fb2b22b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
--
diff --git 
a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
 
b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
index 2f7b34f..020a3c9 100644
--- 
a/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
+++ 
b/phoenix-core/src/main/java/org/apache/phoenix/coprocessor/MetaDataEndpointImpl.java
@@ -17,8 +17,6 @@
  */
 package org.apache.phoenix.coprocessor;
 
-import static com.google.common.collect.Lists.newArrayList;
-import static org.apache.hadoop.hbase.filter.CompareFilter.CompareOp.EQUAL;
 import static org.apache.phoenix.jdbc.PhoenixDatabaseMetaData.ARRAY_SIZE_BYTES;
 import static 
org.apache.phoenix.jdbc.PhoenixDatabaseMetaData.COLUMN_COUNT_BYTES;
 import static 
org.apache.phoenix.jdbc.PhoenixDatabaseMetaData.COLUMN_NAME_INDEX;
@@ -73,15 +71,18 @@ import org.apache.hadoop.hbase.KeyValue;
 import org.apache.hadoop.hbase.KeyValue.Type;
 import org.apache.hadoop.hbase.client.Delete;
 import org.apache.hadoop.hbase.client.Get;
+import org.apache.hadoop.hbase.client.HTableInterface;
+import org.apache.hadoop.hbase.client.HTablePool;
 import org.apache.hadoop.hbase.client.Mutation;
 import org.apache.hadoop.hbase.client.Put;
 import org.apache.hadoop.hbase.client.Result;
+import org.apache.hadoop.hbase.client.ResultScanner;
 import org.apache.hadoop.hbase.client.Scan;
 import org.apache.hadoop.hbase.coprocessor.CoprocessorException;
 import org.apache.hadoop.hbase.coprocessor.CoprocessorService;
 import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment;
+import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp;
 import org.apache.hadoop.hbase.filter.Filter;
-import org.apache.hadoop.hbase.filter.FilterBase;
 import org.apache.hadoop.hbase.filter.FilterList;
 import org.apache.hadoop.hbase.filter.FirstKeyOnlyFilter;
 import org.apache.hadoop.hbase.filter.SingleColumnValueFilter;
@@ -857,12 +858,13 @@ public class MetaDataEndpointImpl extends 
MetaDataProtocol implements Coprocesso
 locks.add(rowLock);
 }
 
-private static final byte[] PHYSICAL_TABLE_BYTES = new byte[] 
{PTable.LinkType.PHYSICAL_TABLE.getSerializedValue()};
+protected static final byte[] PHYSICAL_TABLE_BYTES = new byte[] 
{PTable.LinkType.PHYSICAL_TABLE.getSerializedValue()};
 /**
  * @param tableName parent table's name
  * @return true if there exist a table that use this table as their base 
table.
  * TODO: should we pass a timestamp here?
  */
+@SuppressWarnings("deprecation")
 private boolean hasViews(HRegion region, byte[] tenantId, PTable table) 
throws IOException {
 byte[] schemaName = table.getSchemaName().getBytes();
 byte[] tableName = table.getTableName().getBytes();
@@ -877,21 +879,36 @@ public class MetaDataEndpointImpl extends 
MetaDataProtocol implements Coprocesso
 scan.setStartRow(startRow);
 scan.setStopRow(stopRow);
 }
-SingleColumnValueFilter linkFilter = new 
SingleColumnValueFilter(TABLE_FAMILY_BYTES, LINK_TYPE_BYTES, EQUAL, 
PHYSICAL_TABLE_BYTES);
+SingleColumnValueFilter linkFilter = new 
SingleColumnValueFilter(TABLE_FAMILY_BYTES, LINK_TYPE_BYTES, CompareOp.EQUAL, 
PHYSICAL_TABLE_BYTES);
 linkFilter.setFilterIfMissing(true);
 byte[] suffix = ByteUtil.concat(QueryConstants.SEPARATOR_BYTE_ARRAY, 
SchemaUtil.getTableNameAsBytes(schemaName, tableName));
 SuffixFilter rowFilter = new SuffixFilter(suffix);
 Filter filter = new FilterList(linkFilter, rowFilter);
 scan.setFilter(filter);
 scan.addColumn(TABLE_FAMILY_BYTES, LINK_TYPE_BYTES);
-RegionScanner scanner = region.getScanner(scan);
+//