Thanks Mike & Ryan. Now I can finally see my 5KB messages. However I am running into the following error.
OpenJDK 64-Bit Server VM warning: INFO: os::commit_memory(0x0000000734700000, 530579456, 0) failed; error='Cannot allocate memory' (errno=12) # There is insufficient memory for the Java Runtime Environment to continue. # Native memory allocation (mmap) failed to map 530579456 bytes for committing reserved memory. # An error report file with more information is saved as: I am running spark driver program in the client mode on a standalone cluster using spark 2.1.1. When things happen like this I wonder which memory I need to increase and how? Should I increase the driver JVM memory or executor JVM memory? On Tue, May 16, 2017 at 4:34 PM, Michael Armbrust <mich...@databricks.com> wrote: > I mean the actual kafka client: > > <dependency> > <groupId>org.apache.kafka</groupId> > <artifactId>kafka-clients</artifactId> > <version>0.10.0.1</version> > </dependency> > > > On Tue, May 16, 2017 at 4:29 PM, kant kodali <kanth...@gmail.com> wrote: > >> Hi Michael, >> >> Thanks for the catch. I assume you meant >> *spark-streaming-kafka-0-10_2.11-2.1.0.jar* >> >> I add this in all spark machines under SPARK_HOME/jars. >> >> Still same error seems to persist. Is that the right jar or is there >> anything else I need to add? >> >> Thanks! >> >> >> >> On Tue, May 16, 2017 at 1:40 PM, Michael Armbrust <mich...@databricks.com >> > wrote: >> >>> Looks like you are missing the kafka dependency. >>> >>> On Tue, May 16, 2017 at 1:04 PM, kant kodali <kanth...@gmail.com> wrote: >>> >>>> Looks like I am getting the following runtime exception. I am using >>>> Spark 2.1.0 and the following jars >>>> >>>> *spark-sql_2.11-2.1.0.jar* >>>> >>>> *spark-sql-kafka-0-10_2.11-2.1.0.jar* >>>> >>>> *spark-streaming_2.11-2.1.0.jar* >>>> >>>> >>>> Exception in thread "stream execution thread for [id = >>>> fcfe1fa6-dab3-4769-9e15-e074af622cc1, runId = >>>> 7c54940a-e453-41de-b256-049b539b59b1]" >>>> >>>> java.lang.NoClassDefFoundError: >>>> org/apache/kafka/common/serialization/ByteArrayDeserializer >>>> at >>>> org.apache.spark.sql.kafka010.KafkaSourceProvider.createSource(KafkaSourceProvider.scala:74) >>>> at >>>> org.apache.spark.sql.execution.datasources.DataSource.createSource(DataSource.scala:245) >>>> at >>>> org.apache.spark.sql.execution.streaming.StreamExecution$$anonfun$logicalPlan$1.applyOrElse(StreamExecution.scala:127) >>>> at >>>> org.apache.spark.sql.execution.streaming.StreamExecution$$anonfun$logicalPlan$1.applyOrElse(StreamExecution.scala:123) >>>> at >>>> org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$3.apply(TreeNode.scala:288) >>>> at >>>> org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$3.apply(TreeNode.scala:288) >>>> at >>>> org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70) >>>> >>>> >>>> On Tue, May 16, 2017 at 10:30 AM, Shixiong(Ryan) Zhu < >>>> shixi...@databricks.com> wrote: >>>> >>>>> The default "startingOffsets" is "latest". If you don't push any data >>>>> after starting the query, it won't fetch anything. You can set it to >>>>> "earliest" like ".option("startingOffsets", "earliest")" to start the >>>>> stream from the beginning. >>>>> >>>>> On Tue, May 16, 2017 at 12:36 AM, kant kodali <kanth...@gmail.com> >>>>> wrote: >>>>> >>>>>> Hi All, >>>>>> >>>>>> I have the following code. >>>>>> >>>>>> val ds = sparkSession.readStream() >>>>>> .format("kafka") >>>>>> .option("kafka.bootstrap.servers",bootstrapServers)) >>>>>> .option("subscribe", topicName) >>>>>> .option("checkpointLocation", hdfsCheckPointDir) >>>>>> .load(); >>>>>> >>>>>> val ds1 = ds.select($"value") >>>>>> val query = >>>>>> ds1.writeStream.outputMode("append").format("console").start() >>>>>> query.awaitTermination() >>>>>> >>>>>> There are no errors when I execute this code however I don't see any >>>>>> data being printed out to console? When I run my standalone test Kafka >>>>>> consumer jar I can see that it is receiving messages. so I am not sure >>>>>> what >>>>>> is going on with above code? any ideas? >>>>>> >>>>>> Thanks! >>>>>> >>>>> >>>>> >>>> >>> >> >