I came across these APIs in one the scala tutorials over the net. On Mon, Sep 15, 2014 at 10:14 PM, Mohit Jaggi <mohitja...@gmail.com> wrote:
> But the above APIs are not for HDFS. > > On Mon, Sep 15, 2014 at 9:40 AM, rapelly kartheek <kartheek.m...@gmail.com > > wrote: > >> Yes. I have HDFS. My cluster has 5 nodes. When I run the above commands, >> I see that the file gets created in the master node. But, there wont be any >> data written to it. >> >> >> On Mon, Sep 15, 2014 at 10:06 PM, Mohit Jaggi <mohitja...@gmail.com> >> wrote: >> >>> Is this code running in an executor? You need to make sure the file is >>> accessible on ALL executors. One way to do that is to use a distributed >>> filesystem like HDFS or GlusterFS. >>> >>> On Mon, Sep 15, 2014 at 8:51 AM, rapelly kartheek < >>> kartheek.m...@gmail.com> wrote: >>> >>>> Hi >>>> >>>> I am trying to perform some read/write file operations in spark. >>>> Somehow I am neither able to write to a file nor read. >>>> >>>> import java.io._ >>>> >>>> val writer = new PrintWriter(new File("test.txt" )) >>>> >>>> writer.write("Hello Scala") >>>> >>>> >>>> Can someone please tell me how to perform file I/O in spark. >>>> >>>> >>> >> >