Hi Ilya ,
Thanks for the solution , it worked .
But can you please explain why  overwrite = true is required in case when i
run spark in master slave configuration .

On Wed, Mar 13, 2019 at 8:26 PM Ilya Kasnacheev <[email protected]>
wrote:

> Hello!
>
> Please try savePairs(rdd, true).
>
> Hope it helps!
> --
> Ilya Kasnacheev
>
>
> ср, 13 мар. 2019 г. в 17:41, Harshal Patil <[email protected]>:
>
>> Hi ,
>> I  am using SPARK 2.3.0 and Ignite 2.7.0 . I have enabled POSTGRES as
>> persistent store through gridgain automatic RDBMS integration .I have
>> enabled write through cache .  I could see data being persisted in POSTGRES
>> when I am running SPARK in standalone mode , with
>>
>> *val *conf = *new *SparkConf()
>>
>> conf.setMaster("local[*]")
>>
>>
>> *But* when I have master configured for spark like,
>>
>>
>> conf.setMaster("spark://harshal-patil.local:7077")
>>
>>
>> my data is not getting persisted in postgres , but i can see cache is
>> updated .
>>
>>
>> I am doing operation - ic.fromCache("RoleCache").savePairs(rdd)
>>
>>
>> Please help me understand what could be going wrong .
>>
>>
>> Regards ,
>>
>> Harshal
>>
>>
>>
>>

Reply via email to