Ram Mettu created KUDU-1581:
-------------------------------
Summary: Kudu-Spark read failure when the Kudu table contains
BINARY column
Key: KUDU-1581
URL: https://issues.apache.org/jira/browse/KUDU-1581
Project: Kudu
Issue Type: Bug
Components: client
Affects Versions: 0.10.0
Reporter: Ram Mettu
Assignee: Ram Mettu
Using kudu-spark, create a Spark dataframe for a Kudu table containing BINARY
column, any action fails to serialize.
Steps to reproduce:
1. Create kudu table with binary column(s)
2. Populate table with data
3. Create Spark Dataframe and perform an action
val data = sqlContext.read.options(Map("kudu.master" -> masterAddress,
"kudu.table" -> "test")).kudu
data.show()
Results in an error
org.apache.spark.SparkException: Job aborted due to stage failure: Task 0.0 in
stage 1.0 (TID 1) had a not serializable result: java.nio.HeapByteBuffer
Serialization stack:
- object not serializable (class: java.nio.HeapByteBuffer, value:
java.nio.HeapByteBuffer[pos=677 lim=682 cap=727])
- element of array (index: 8)
- array (class [Ljava.lang.Object;, size 9)
- field (class:
org.apache.spark.sql.catalyst.expressions.GenericInternalRow, name: values,
type: class [Ljava.lang.Object;)
- object (class
org.apache.spark.sql.catalyst.expressions.GenericInternalRow,
[0,0,0,0.0,0,false,0,0.0,java.nio.HeapByteBuffer[pos=677 lim=682 cap=727]])
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)