[jira] [Commented] (SPARK-3807) SparkSql does not work for tables created using custom serde

2014-10-16 Thread Apache Spark (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-3807?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14173430#comment-14173430
 ] 

Apache Spark commented on SPARK-3807:
-

User 'adrian-wang' has created a pull request for this issue:
https://github.com/apache/spark/pull/2821

 SparkSql does not work for tables created using custom serde
 

 Key: SPARK-3807
 URL: https://issues.apache.org/jira/browse/SPARK-3807
 Project: Spark
  Issue Type: Bug
  Components: SQL
Affects Versions: 1.1.0
Reporter: chirag aggarwal
 Fix For: 1.1.1, 1.2.0


 SparkSql crashes on selecting tables using custom serde. 
 Example:
 
 CREATE EXTERNAL TABLE table_name PARTITIONED BY ( a int) ROW FORMAT 'SERDE 
 org.apache.hadoop.hive.serde2.thrift.ThriftDeserializer with 
 serdeproperties(serialization.format=org.apache.thrift.protocol.TBinaryProtocol,serialization.class=ser_class)
  STORED AS SEQUENCEFILE;
 The following exception is seen on running a query like 'select * from 
 table_name limit 1': 
 ERROR CliDriver: org.apache.hadoop.hive.serde2.SerDeException: 
 java.lang.NullPointerException 
 at 
 org.apache.hadoop.hive.serde2.thrift.ThriftDeserializer.initialize(ThriftDeserializer.java:68)
  
 at 
 org.apache.hadoop.hive.ql.plan.TableDesc.getDeserializer(TableDesc.java:80) 
 at 
 org.apache.spark.sql.hive.execution.HiveTableScan.addColumnMetadataToConf(HiveTableScan.scala:86)
  
 at 
 org.apache.spark.sql.hive.execution.HiveTableScan.init(HiveTableScan.scala:100)
  
 at 
 org.apache.spark.sql.hive.HiveStrategies$HiveTableScans$$anonfun$14.apply(HiveStrategies.scala:188)
  
 at 
 org.apache.spark.sql.hive.HiveStrategies$HiveTableScans$$anonfun$14.apply(HiveStrategies.scala:188)
  
 at 
 org.apache.spark.sql.SQLContext$SparkPlanner.pruneFilterProject(SQLContext.scala:364)
  
 at 
 org.apache.spark.sql.hive.HiveStrategies$HiveTableScans$.apply(HiveStrategies.scala:184)
  
 at 
 org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
  
 at 
 org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
  
 at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) 
 at 
 org.apache.spark.sql.catalyst.planning.QueryPlanner.apply(QueryPlanner.scala:59)
  
 at 
 org.apache.spark.sql.catalyst.planning.QueryPlanner.planLater(QueryPlanner.scala:54)
  
 at 
 org.apache.spark.sql.execution.SparkStrategies$BasicOperators$.apply(SparkStrategies.scala:280)
  
 at 
 org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
  
 at 
 org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
  
 at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) 
 at 
 org.apache.spark.sql.catalyst.planning.QueryPlanner.apply(QueryPlanner.scala:59)
  
 at 
 org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan$lzycompute(SQLContext.scala:402)
  
 at 
 org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan(SQLContext.scala:400)
  
 at 
 org.apache.spark.sql.SQLContext$QueryExecution.executedPlan$lzycompute(SQLContext.scala:406)
  
 at 
 org.apache.spark.sql.SQLContext$QueryExecution.executedPlan(SQLContext.scala:406)
  
 at 
 org.apache.spark.sql.hive.HiveContext$QueryExecution.stringResult(HiveContext.scala:406)
  
 at 
 org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.run(SparkSQLDriver.scala:59)
  
 at 
 org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:291)
  
 at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:413) 
 at 
 org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:226)
  
 at 
 org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(SparkSQLCLIDriver.scala)
  
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 
 at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) 
 at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) 
 at java.lang.reflect.Method.invoke(Unknown Source) 
 at org.apache.spark.deploy.SparkSubmit$.launch(SparkSubmit.scala:328) 
 at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:75) 
 at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) 
 Caused by: java.lang.NullPointerException
 After fixing this issue, when some columns in the table were referred in the 
 query, sparksql could not resolve those references.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-3807) SparkSql does not work for tables created using custom serde

2014-10-06 Thread chirag aggarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-3807?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14160197#comment-14160197
 ] 

chirag aggarwal commented on SPARK-3807:


Pull request for this issue:
https://github.com/apache/spark/pull/2674

 SparkSql does not work for tables created using custom serde
 

 Key: SPARK-3807
 URL: https://issues.apache.org/jira/browse/SPARK-3807
 Project: Spark
  Issue Type: Bug
  Components: SQL
Affects Versions: 1.1.0
Reporter: chirag aggarwal
 Fix For: 1.1.1


 SparkSql crashes on selecting tables using custom serde. 
 The following exception is seen on running a query like 'select * from 
 table_name limit 1': 
 ERROR CliDriver: org.apache.hadoop.hive.serde2.SerDeException: 
 java.lang.NullPointerException 
 at 
 org.apache.hadoop.hive.serde2.thrift.ThriftDeserializer.initialize(ThriftDeserializer.java:68)
  
 at 
 org.apache.hadoop.hive.ql.plan.TableDesc.getDeserializer(TableDesc.java:80) 
 at 
 org.apache.spark.sql.hive.execution.HiveTableScan.addColumnMetadataToConf(HiveTableScan.scala:86)
  
 at 
 org.apache.spark.sql.hive.execution.HiveTableScan.init(HiveTableScan.scala:100)
  
 at 
 org.apache.spark.sql.hive.HiveStrategies$HiveTableScans$$anonfun$14.apply(HiveStrategies.scala:188)
  
 at 
 org.apache.spark.sql.hive.HiveStrategies$HiveTableScans$$anonfun$14.apply(HiveStrategies.scala:188)
  
 at 
 org.apache.spark.sql.SQLContext$SparkPlanner.pruneFilterProject(SQLContext.scala:364)
  
 at 
 org.apache.spark.sql.hive.HiveStrategies$HiveTableScans$.apply(HiveStrategies.scala:184)
  
 at 
 org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
  
 at 
 org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
  
 at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) 
 at 
 org.apache.spark.sql.catalyst.planning.QueryPlanner.apply(QueryPlanner.scala:59)
  
 at 
 org.apache.spark.sql.catalyst.planning.QueryPlanner.planLater(QueryPlanner.scala:54)
  
 at 
 org.apache.spark.sql.execution.SparkStrategies$BasicOperators$.apply(SparkStrategies.scala:280)
  
 at 
 org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
  
 at 
 org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
  
 at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) 
 at 
 org.apache.spark.sql.catalyst.planning.QueryPlanner.apply(QueryPlanner.scala:59)
  
 at 
 org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan$lzycompute(SQLContext.scala:402)
  
 at 
 org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan(SQLContext.scala:400)
  
 at 
 org.apache.spark.sql.SQLContext$QueryExecution.executedPlan$lzycompute(SQLContext.scala:406)
  
 at 
 org.apache.spark.sql.SQLContext$QueryExecution.executedPlan(SQLContext.scala:406)
  
 at 
 org.apache.spark.sql.hive.HiveContext$QueryExecution.stringResult(HiveContext.scala:406)
  
 at 
 org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.run(SparkSQLDriver.scala:59)
  
 at 
 org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:291)
  
 at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:413) 
 at 
 org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:226)
  
 at 
 org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(SparkSQLCLIDriver.scala)
  
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 
 at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) 
 at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) 
 at java.lang.reflect.Method.invoke(Unknown Source) 
 at org.apache.spark.deploy.SparkSubmit$.launch(SparkSubmit.scala:328) 
 at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:75) 
 at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) 
 Caused by: java.lang.NullPointerException



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-3807) SparkSql does not work for tables created using custom serde

2014-10-06 Thread Apache Spark (JIRA)

[ 
https://issues.apache.org/jira/browse/SPARK-3807?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14160200#comment-14160200
 ] 

Apache Spark commented on SPARK-3807:
-

User 'chiragaggarwal' has created a pull request for this issue:
https://github.com/apache/spark/pull/2674

 SparkSql does not work for tables created using custom serde
 

 Key: SPARK-3807
 URL: https://issues.apache.org/jira/browse/SPARK-3807
 Project: Spark
  Issue Type: Bug
  Components: SQL
Affects Versions: 1.1.0
Reporter: chirag aggarwal
 Fix For: 1.1.1


 SparkSql crashes on selecting tables using custom serde. 
 The following exception is seen on running a query like 'select * from 
 table_name limit 1': 
 ERROR CliDriver: org.apache.hadoop.hive.serde2.SerDeException: 
 java.lang.NullPointerException 
 at 
 org.apache.hadoop.hive.serde2.thrift.ThriftDeserializer.initialize(ThriftDeserializer.java:68)
  
 at 
 org.apache.hadoop.hive.ql.plan.TableDesc.getDeserializer(TableDesc.java:80) 
 at 
 org.apache.spark.sql.hive.execution.HiveTableScan.addColumnMetadataToConf(HiveTableScan.scala:86)
  
 at 
 org.apache.spark.sql.hive.execution.HiveTableScan.init(HiveTableScan.scala:100)
  
 at 
 org.apache.spark.sql.hive.HiveStrategies$HiveTableScans$$anonfun$14.apply(HiveStrategies.scala:188)
  
 at 
 org.apache.spark.sql.hive.HiveStrategies$HiveTableScans$$anonfun$14.apply(HiveStrategies.scala:188)
  
 at 
 org.apache.spark.sql.SQLContext$SparkPlanner.pruneFilterProject(SQLContext.scala:364)
  
 at 
 org.apache.spark.sql.hive.HiveStrategies$HiveTableScans$.apply(HiveStrategies.scala:184)
  
 at 
 org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
  
 at 
 org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
  
 at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) 
 at 
 org.apache.spark.sql.catalyst.planning.QueryPlanner.apply(QueryPlanner.scala:59)
  
 at 
 org.apache.spark.sql.catalyst.planning.QueryPlanner.planLater(QueryPlanner.scala:54)
  
 at 
 org.apache.spark.sql.execution.SparkStrategies$BasicOperators$.apply(SparkStrategies.scala:280)
  
 at 
 org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
  
 at 
 org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
  
 at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) 
 at 
 org.apache.spark.sql.catalyst.planning.QueryPlanner.apply(QueryPlanner.scala:59)
  
 at 
 org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan$lzycompute(SQLContext.scala:402)
  
 at 
 org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan(SQLContext.scala:400)
  
 at 
 org.apache.spark.sql.SQLContext$QueryExecution.executedPlan$lzycompute(SQLContext.scala:406)
  
 at 
 org.apache.spark.sql.SQLContext$QueryExecution.executedPlan(SQLContext.scala:406)
  
 at 
 org.apache.spark.sql.hive.HiveContext$QueryExecution.stringResult(HiveContext.scala:406)
  
 at 
 org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.run(SparkSQLDriver.scala:59)
  
 at 
 org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:291)
  
 at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:413) 
 at 
 org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:226)
  
 at 
 org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(SparkSQLCLIDriver.scala)
  
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 
 at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) 
 at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) 
 at java.lang.reflect.Method.invoke(Unknown Source) 
 at org.apache.spark.deploy.SparkSubmit$.launch(SparkSubmit.scala:328) 
 at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:75) 
 at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) 
 Caused by: java.lang.NullPointerException



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org