But i just want to update rdd , by  appending  unique message ID  with
each  element of RDD , which should be automatically(m++ ..) updated every
time  a new element comes to rdd .

On Mon, Jun 22, 2015 at 7:05 AM, Michal Čizmazia <mici...@gmail.com> wrote:

> StreamingContext.sparkContext()
>
> On 21 June 2015 at 21:32, Will Briggs <wrbri...@gmail.com> wrote:
>
>> It sounds like accumulators are not necessary in Spark Streaming - see
>> this post (
>> http://apache-spark-user-list.1001560.n3.nabble.com/Shared-variable-in-Spark-Streaming-td11762.html)
>> for more details.
>>
>>
>> On June 21, 2015, at 7:31 PM, anshu shukla <anshushuk...@gmail.com>
>> wrote:
>>
>>
>> In spark Streaming ,Since we are already having Streaming context ,
>>  which does not allows us to have accumulators .We have to get sparkContext
>>  for initializing accumulator value .
>> But  having 2 spark context will not serve the problem .
>>
>> Please Help !!
>>
>> --
>> Thanks & Regards,
>> Anshu Shukla
>>
>
>


-- 
Thanks & Regards,
Anshu Shukla

Reply via email to