Yes. I have HDFS. My cluster has 5 nodes. When I run the above commands, I see that the file gets created in the master node. But, there wont be any data written to it.
On Mon, Sep 15, 2014 at 10:06 PM, Mohit Jaggi <mohitja...@gmail.com> wrote: > Is this code running in an executor? You need to make sure the file is > accessible on ALL executors. One way to do that is to use a distributed > filesystem like HDFS or GlusterFS. > > On Mon, Sep 15, 2014 at 8:51 AM, rapelly kartheek <kartheek.m...@gmail.com > > wrote: > >> Hi >> >> I am trying to perform some read/write file operations in spark. Somehow >> I am neither able to write to a file nor read. >> >> import java.io._ >> >> val writer = new PrintWriter(new File("test.txt" )) >> >> writer.write("Hello Scala") >> >> >> Can someone please tell me how to perform file I/O in spark. >> >> >