[
https://issues.apache.org/jira/browse/PHOENIX-2889?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Ian Hellstrom updated PHOENIX-2889:
-----------------------------------
Description:
When using {{phoenixTableAsDataFrame}} on a table with auto-capitalized
qualifiers where the user has erroneously specified these with lower caps, no
exception is returned. Ideally, an
{{org.apache.phoenix.schema.ColumnNotFoundException}} is thrown but instead
lines like the following show up in the log
{code}
INFO RpcRetryingCaller: Call exception, tries=10, retries=35, started=48168 ms
ago, cancelled=false, msg=
{code}
A minimal example:
{code}
CREATE TABLE test (foo INT, bar VARCHAR);
UPSERT INTO test VALUES (1, 'hello');
UPSERT INTO test VALUES (2, 'bye');
{code}
In Spark (shell):
{code}
import org.apache.hadoop.conf.Configuration
import org.apache.hadoop.hbase.HBaseConfiguration
import org.apache.phoenix.spark._
@transient lazy val hadoopConf: Configuration = new Configuration()
@transient lazy val hbConf: Configuration =
HBaseConfiguration.create(hadoopConf)
val df = sqlContext.phoenixTableAsDataFrame("TEST", Array("foo", "bar"), conf =
hbConf)
{code}
was:
When using {{phoenixTableAsDataFrame}} on a table with auto-capitalized
qualifiers where the user has erroneously specified these with lower caps, no
exception is returned. Ideally, a
{{org.apache.phoenix.schema.ColumnNotFoundException}} is thrown but instead
lines like the following show up in the log
{code}
INFO RpcRetryingCaller: Call exception, tries=10, retries=35, started=48168 ms
ago, cancelled=false, msg=
{code}
A minimal example:
{code}
CREATE TABLE test (foo INT, bar VARCHAR);
UPSERT INTO test VALUES (1, 'hello');
UPSERT INTO test VALUES (2, 'bye');
{code}
In Spark (shell):
{code}
import org.apache.hadoop.conf.Configuration
import org.apache.hadoop.hbase.HBaseConfiguration
import org.apache.phoenix.spark._
@transient lazy val hadoopConf: Configuration = new Configuration()
@transient lazy val hbConf: Configuration =
HBaseConfiguration.create(hadoopConf)
val df = sqlContext.phoenixTableAsDataFrame("TEST", Array("foo", "bar"), conf =
hbConf)
{code}
> No exception in Phoenix-Spark when column does not exist
> --------------------------------------------------------
>
> Key: PHOENIX-2889
> URL: https://issues.apache.org/jira/browse/PHOENIX-2889
> Project: Phoenix
> Issue Type: Bug
> Affects Versions: 4.4.0
> Environment: Phoenix: 4.4
> HBase: 1.1.2.2.3.2.0-2950 (Hortonworks HDP)
> Spark: 1.4.1 (compiled with Scala 2.10)
> Reporter: Ian Hellstrom
> Labels: spark
>
> When using {{phoenixTableAsDataFrame}} on a table with auto-capitalized
> qualifiers where the user has erroneously specified these with lower caps, no
> exception is returned. Ideally, an
> {{org.apache.phoenix.schema.ColumnNotFoundException}} is thrown but instead
> lines like the following show up in the log
> {code}
> INFO RpcRetryingCaller: Call exception, tries=10, retries=35, started=48168
> ms ago, cancelled=false, msg=
> {code}
> A minimal example:
> {code}
> CREATE TABLE test (foo INT, bar VARCHAR);
> UPSERT INTO test VALUES (1, 'hello');
> UPSERT INTO test VALUES (2, 'bye');
> {code}
> In Spark (shell):
> {code}
> import org.apache.hadoop.conf.Configuration
> import org.apache.hadoop.hbase.HBaseConfiguration
> import org.apache.phoenix.spark._
> @transient lazy val hadoopConf: Configuration = new Configuration()
> @transient lazy val hbConf: Configuration =
> HBaseConfiguration.create(hadoopConf)
> val df = sqlContext.phoenixTableAsDataFrame("TEST", Array("foo", "bar"), conf
> = hbConf)
> {code}
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)