This works fine with Kafka and Flink. However when I do it with spark , new
line feed gets removed.

On Tue, Feb 16, 2016 at 4:29 PM, UMESH CHAUDHARY <umesh9...@gmail.com>
wrote:

> Try to print RDD before writing to validate that you are getting '\n' from
> Kafka.
>
> On Tue, Feb 16, 2016 at 4:19 PM, Ashutosh Kumar <kmr.ashutos...@gmail.com>
> wrote:
>
>> Hi Chandeep,
>> Thanks for response. Issue is the new line feed is lost. All records
>> appear in one line only.
>>
>> Thanks
>> Ashutosh
>>
>> On Tue, Feb 16, 2016 at 3:26 PM, Chandeep Singh <c...@chandeep.com> wrote:
>>
>>> !rdd.isEmpty() should work but an alternative could be rdd.take(1) != 0
>>>
>>> On Feb 16, 2016, at 9:33 AM, Ashutosh Kumar <kmr.ashutos...@gmail.com>
>>> wrote:
>>>
>>> I am getting multiple empty files for streaming output for each interval.
>>> To Avoid this I tried
>>>
>>>      kStream.foreachRDD(new VoidFunction2<JavaRDD<String>,Time>(){
>>>
>>>
>>>
>>>
>>>
>>> *public void call(JavaRDD<String> rdd,Time time) throws Exception {
>>>                 if(!rdd.isEmpty()){
>>> rdd.saveAsTextFile("filename_"+time.milliseconds()+".csv");
>>>     }                }*
>>> This prevents writing of empty files. However this appends line after
>>> one another by removing new lines. All lines are merged.
>>> How do I retain my new line?
>>>
>>> Thanks
>>>
>>>
>>>
>>
>

Reply via email to