But i just want to update rdd , by appending unique message ID with each element of RDD , which should be automatically(m++ ..) updated every time a new element comes to rdd .
On Mon, Jun 22, 2015 at 7:05 AM, Michal Čizmazia <mici...@gmail.com> wrote: > StreamingContext.sparkContext() > > On 21 June 2015 at 21:32, Will Briggs <wrbri...@gmail.com> wrote: > >> It sounds like accumulators are not necessary in Spark Streaming - see >> this post ( >> http://apache-spark-user-list.1001560.n3.nabble.com/Shared-variable-in-Spark-Streaming-td11762.html) >> for more details. >> >> >> On June 21, 2015, at 7:31 PM, anshu shukla <anshushuk...@gmail.com> >> wrote: >> >> >> In spark Streaming ,Since we are already having Streaming context , >> which does not allows us to have accumulators .We have to get sparkContext >> for initializing accumulator value . >> But having 2 spark context will not serve the problem . >> >> Please Help !! >> >> -- >> Thanks & Regards, >> Anshu Shukla >> > > -- Thanks & Regards, Anshu Shukla