[
https://issues.apache.org/jira/browse/CARBONDATA-1772?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16265581#comment-16265581
]
QiangCai commented on CARBONDATA-1772:
--------------------------------------
[~Bjangir]
This issue was fixed by pull request 1530.
https://github.com/apache/carbondata/pull/1530
> [Streaming]carbon StreamWriter task is failled with ClassCastException
> ----------------------------------------------------------------------
>
> Key: CARBONDATA-1772
> URL: https://issues.apache.org/jira/browse/CARBONDATA-1772
> Project: CarbonData
> Issue Type: Bug
> Affects Versions: 1.3.0
> Reporter: Babulal
> Attachments: streamingLog.log
>
>
> Run below Seq of commands in spark Shell ( bin/spark-shell --jars
> /opt/carbon/carbondata_2.11-1.3.0-SNAPSHOT-shade-hadoop2.7.2.jar --master
> yarn-client --executor-memory 1G --executor-cores 2 --driver-memory 1G
> )
> // carbon is SparkSession with CarbonStateBuilder
> carbon.sql("create table stable (value String,count String) STORED BY
> 'carbondata' TBLPROPERTIES ('streaming' = 'true')")
> val lines = carbon.readStream.format("socket") .option("host", "localhost")
> .option("port", 9999) .load()
> val words = lines.as[String].flatMap(_.split(" "))
> val wordCounts = words.groupBy("value").count()
> val carbonTable = CarbonEnv.getCarbonTable(Some("default"), "stable")(carbon)
> val tablePath =
> CarbonStorePath.getCarbonTablePath(carbonTable.getAbsoluteTableIdentifier)
> val qry =
> wordCounts.writeStream.format("carbondata").outputMode("complete").trigger(ProcessingTime("1
> seconds")).option("tablePath",
> tablePath.getPath).option("checkpointLocation",
> tablePath.getStreamingCheckpointDir).option("tableName","stable").start()
> scala> qry.awaitTermination()
> Now in another window run below command
> root@master ~ # nc -lk 9999
> babu
> Check SparkShell
> Stage 1:> (0 + 6) /
> 200]17/11/19 17:59:57 WARN TaskSetManager: Lost task 2.0 in stage 1.0 (TID 3,
> slave1, executor 2): org.apache.carbondata.streaming.CarbonStreamException:
> Task failed while writing rows
> at
> org.apache.spark.sql.execution.streaming.CarbonAppendableStreamSink$.writeDataFileTask(CarbonAppendableStreamSink.scala:286)
> at
> org.apache.spark.sql.execution.streaming.CarbonAppendableStreamSink$$anonfun$writeDataFileJob$1$$anonfun$apply$mcV$sp$1.apply(CarbonAppendableStreamSink.scala:192)
> at
> org.apache.spark.sql.execution.streaming.CarbonAppendableStreamSink$$anonfun$writeDataFileJob$1$$anonfun$apply$mcV$sp$1.apply(CarbonAppendableStreamSink.scala:191)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:87)
> at org.apache.spark.scheduler.Task.run(Task.scala:99)
> at
> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> at java.lang.Thread.run(Thread.java:748)
> Caused by: java.lang.ClassCastException: cannot assign instance of
> scala.collection.immutable.List$SerializationProxy to field
> scala.collection.convert.Wrappers$SeqWrapper.underlying of type
> scala.collection.Seq in instance of
> scala.collection.convert.Wrappers$SeqWrapper
> at
> java.io.ObjectStreamClass$FieldReflector.setObjFieldValues(ObjectStreamClass.java:2133)
> at
> java.io.ObjectStreamClass.setObjFieldValues(ObjectStreamClass.java:1305)
> at
> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2251)
> at
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2169)
> at
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2027)
> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1535)
> at
> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2245)
> at
> java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2169)
> at
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2027)
> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1535)
> at java.io.ObjectInputStream.readObject(ObjectInputStream.java:422)
> at
> org.apache.carbondata.hadoop.util.ObjectSerializationUtil.convertStringToObject(ObjectSerializationUtil.java:99)
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)